Introducing Graph Smoothness Loss for Training Deep Learning Architectures | IEEE Conference Publication | IEEE Xplore

Introducing Graph Smoothness Loss for Training Deep Learning Architectures


Abstract:

We introduce a novel loss function for training deep learning architectures to perform classification. It consists in minimizing the smoothness of label signals on simila...Show More

Abstract:

We introduce a novel loss function for training deep learning architectures to perform classification. It consists in minimizing the smoothness of label signals on similarity graphs built at the output of the architecture. Equivalently, it can be seen as maximizing the distances between the network function images of training inputs from distinct classes. As such, only distances between pairs of examples in distinct classes are taken into account in the process, and the training does not prevent inputs from the same class to be mapped to distant locations in the output domain. We show that this loss leads to similar performance in classification as architectures trained using the classical cross-entropy, while offering interesting degrees of freedom and properties. We also demonstrate the interest of the proposed loss to increase robustness of trained architectures to deviations of the inputs.
Date of Conference: 02-05 June 2019
Date Added to IEEE Xplore: 08 July 2019
ISBN Information:
Conference Location: Minneapolis, MN, USA

Contact IEEE to Subscribe

References

References is not available for this document.