Conference paper Open Access

Subclass deep neural networks: re-enabling neglected classes in deep network training for multimedia classification

Gkalelis, Nikolaos; Mezaris, Vasileios

During minibatch gradient-based optimization, the contribution of observations to the updating of the deep neural network's (DNN's) weights for enhancing the discrimination of certain classes can be small, despite the fact that these classes may still have a large generalization error. This happens, for instance, due to overfitting, i.e. to classes whose error in the training set is negligible, or simply when the contributions of the misclassified observations to the updating of the weights associated with these classes cancel out. To alleviate this problem, a new criterion for identifying the so-called "neglected" classes during the training of DNNs, i.e. the classes which stop to optimize early in the training procedure, is proposed. Moreover, based on this criterion a novel cost function is proposed, that extends the cross-entropy loss using subclass partitions for boosting the generalization performance of the neglected classes. In this way, the network is guided to emphasize the extraction of features that are discriminant for the classes that are prone to being neglected during the optimization procedure. The proposed framework can be easily applied to improve the performance of various DNN architectures. Experiments on several publicly available benchmarks including, the large-scale YouTube-8M (YT8M) video dataset, show the efficacy of the proposed method. Source code is made publicly available at: https://github.com/bmezaris/subclass_deep_neural_networks

Files (2.3 MB)
Name Size
mmm2020_lncs11961_2_preprint.pdf
md5:170972b2ba0a731a23ffeeb1d057117b
2.3 MB Download
73
18
views
downloads
Views 73
Downloads 18
Data volume 40.6 MB
Unique views 71
Unique downloads 17

Share

Cite as