Conference paper Open Access

Subclass deep neural networks: re-enabling neglected classes in deep network training for multimedia classification

Gkalelis, Nikolaos; Mezaris, Vasileios

During minibatch gradient-based optimization, the contribution of observations to the updating of the deep neural network's (DNN's) weights for enhancing the discrimination of certain classes can be small, despite the fact that these classes may still have a large generalization error. This happens, for instance, due to overfitting, i.e. to classes whose error in the training set is negligible, or simply when the contributions of the misclassified observations to the updating of the weights associated with these classes cancel out. To alleviate this problem, a new criterion for identifying the so-called "neglected" classes during the training of DNNs, i.e. the classes which stop to optimize early in the training procedure, is proposed. Moreover, based on this criterion a novel cost function is proposed, that extends the cross-entropy loss using subclass partitions for boosting the generalization performance of the neglected classes. In this way, the network is guided to emphasize the extraction of features that are discriminant for the classes that are prone to being neglected during the optimization procedure. The proposed framework can be easily applied to improve the performance of various DNN architectures. Experiments on several publicly available benchmarks including, the large-scale YouTube-8M (YT8M) video dataset, show the efficacy of the proposed method. Source code is made publicly available at: https://github.com/bmezaris/subclass_deep_neural_networks

Files (2.3 MB)
Name Size
mmm2020_lncs11961_2_preprint.pdf
md5:170972b2ba0a731a23ffeeb1d057117b
2.3 MB Download
29
9
views
downloads
Views 29
Downloads 9
Data volume 20.3 MB
Unique views 27
Unique downloads 8

Share

Cite as