Mixture separability loss in a deep convolutional network for image classification
In machine learning, the cost function is crucial because it measures how good or bad a system is. In image classification, well-known networks only consider modifying the network structures and applying cross-entropy loss at the end of the network. However, using only cross-entropy loss causes a ne...
Saved in:
Published in | IET image processing Vol. 13; no. 1; pp. 135 - 141 |
---|---|
Main Authors | , , , |
Format | Journal Article |
Language | English |
Published |
The Institution of Engineering and Technology
01.01.2019
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | In machine learning, the cost function is crucial because it measures how good or bad a system is. In image classification, well-known networks only consider modifying the network structures and applying cross-entropy loss at the end of the network. However, using only cross-entropy loss causes a network to stop updating weights when all training images are correctly classified. This is the problem of early saturation. This study proposes a novel cost function, called mixture separability loss (MSL), which updates the weights of the network even when most of the training images are accurately predicted. MSL consists of between-class and within-class loss. Between-class loss maximises the differences between inter-class images, whereas within-class loss minimises the similarities between intra-class images. They designed the proposed loss function to attach to different convolutional layers in the network in order to utilise intermediate feature maps. Experiments show that a network with MSL deepens the learning process and obtains promising results with some public datasets, such as Street View House Number, Canadian Institute for Advanced Research, and the authors’ self-collected Inha Computer Vision Lab gender dataset. |
---|---|
ISSN: | 1751-9659 1751-9667 |
DOI: | 10.1049/iet-ipr.2018.5613 |