ETRI-Knowledge Sharing Plaform

KOREAN
논문 검색
Type SCI
Year ~ Keyword

Detail

Conference Paper A Performance Comparison of Loss Functions
Cited 9 time in scopus Share share facebook twitter linkedin kakaostory
Authors
Kwantae Cho, Jong-hyuk Roh, Youngsam Kim, Sangrae Cho
Issue Date
2019-10
Citation
International Conference on Information and Communication Technology Convergence (ICTC) 2019, pp.1146-1151
Publisher
IEEE
Language
English
Type
Conference Paper
DOI
https://dx.doi.org/10.1109/ICTC46691.2019.8939902
Abstract
Generally, the deep neural network learns by way of a loss function, which is an approach to evaluate how well given dataset is predicted on a particular network architecture (or network model). If the prediction deviates too far from real data, a loss function would generate a very large value. Progressively, with the help of some optimization function, the loss function lowers the prediction error by providing the network architecture with information that can control the weights of the network architecture. Thus, the loss functions plays an important role in training the network architecture.Recently, several researchers have studied various loss functions such as Softmax, Modified softmax, Angular softmax, Additive-Margin softmax, Arcface, Center, and Focal losses. In this manuscript, we propose a new and simple loss function that just adds the existing loss functions. In addition, we conduct experiments with the MNIST dataset in order to compare the performance between all loss functions including the proposed and the existing loss functions. Resultingly, the experiments show that the proposed loss function is visibly superior to the ability to classify digit images. The experimental results also indicate that Arcface loss and Additive-Margin loss functions satisfy predefined test accuracy most quickly under two and three dimensional embedding, respectively. The fast learning ability of the both loss functions has the advantage of providing relatively high accuracy even when the number of train data is small.
KSP Keywords
Deep neural network(DNN), Fast learning, High accuracy, Learning ability, MNIST Dataset, Network Architecture, Network model, Performance comparison, Prediction error, Real data, Three dimensional(3D)