Mixture separability loss in a deep convolutional network for image classification

In machine learning, the cost function is crucial because it measures how good or bad a system is. In image classification, well-known networks only consider modifying the network structures and applying cross-entropy loss at the end of the network. However, using only cross-entropy loss causes a ne...

Full description

Saved in:
Bibliographic Details
Published inIET image processing Vol. 13; no. 1; pp. 135 - 141
Main Authors Do, Trung Dung, Jin, Cheng-Bin, Nguyen, Van Huan, Kim, Hakil
Format Journal Article
LanguageEnglish
Published The Institution of Engineering and Technology 01.01.2019
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:In machine learning, the cost function is crucial because it measures how good or bad a system is. In image classification, well-known networks only consider modifying the network structures and applying cross-entropy loss at the end of the network. However, using only cross-entropy loss causes a network to stop updating weights when all training images are correctly classified. This is the problem of early saturation. This study proposes a novel cost function, called mixture separability loss (MSL), which updates the weights of the network even when most of the training images are accurately predicted. MSL consists of between-class and within-class loss. Between-class loss maximises the differences between inter-class images, whereas within-class loss minimises the similarities between intra-class images. They designed the proposed loss function to attach to different convolutional layers in the network in order to utilise intermediate feature maps. Experiments show that a network with MSL deepens the learning process and obtains promising results with some public datasets, such as Street View House Number, Canadian Institute for Advanced Research, and the authors’ self-collected Inha Computer Vision Lab gender dataset.
ISSN:1751-9659
1751-9667
DOI:10.1049/iet-ipr.2018.5613