Vincent Gripon's Homepage

Research and Teaching Blog

Introducing Graph Smoothness Loss for Training Deep Learning Architectures

M. Bontonou, C. Lassance, G. B. Hacene, V. Gripon, J. Tang and A. Ortega, "Introducing Graph Smoothness Loss for Training Deep Learning Architectures," in Data Science Workshop, pp. 160--164, June 2019.

We introduce a novel loss function for training deep learning architectures to perform classification. It consists in minimizing the smoothness of label signals on similarity graphs built at the output of the architecture. Equivalently, it can be seen as maximizing the distances between the network function images of training inputs from distinct classes. As such, only distances between pairs of examples in distinct classes are taken into account in the process, and the training does not prevent inputs from the same class to be mapped to distant locations in the output domain. We show that this loss leads to similar performance in classification as architectures trained using the classical cross-entropy, while offering interesting degrees of freedom and properties. We also demonstrate the interest of the proposed loss to increase robustness of trained architectures to deviations of the inputs.

Download manuscript.

Bibtex
@inproceedings{BonLasHacGriTanOrt20196,
  author = {Myriam Bontonou and Carlos Lassance and
Ghouthi Boukli Hacene and Vincent Gripon and Jian Tang
and Antonio Ortega},
  title = {Introducing Graph Smoothness Loss for
Training Deep Learning Architectures},
  booktitle = {Data Science Workshop},
  year = {2019},
  pages = {160--164},
  month = {June},
}




You are the 1975647th visitor

Vincent Gripon's Homepage