Vincent Gripon's Homepage

Research and Teaching Blog

Budget Restricted Incremental Learning with Pre-Trained Convolutional Neural Networks and Binary Associative Memories

G. B. Hacene, V. Gripon, N. Farrugia, M. Arzel and M. Jezequel, "Budget Restricted Incremental Learning with Pre-Trained Convolutional Neural Networks and Binary Associative Memories," in Proceedings of SIPS, 2017. To appear.

Thanks to their ability to absorb large amounts of data, Convolutional Neural Networks (CNNs) have become state-of-the-art in numerous vision challenges, sometimes even on par with biological vision. They rely on optimisation routines that typically require intensive computational power, thus the question of embedded architectures is a very active field of research. Of particular interest is the problem of incremental learning, where the device adapts to new observations or classes. To tackle this challenging problem, we propose to combine pre-trained CNNs with binary associative memories, using product random sampling as an intermediate between the two methods. The obtained architecture requires significantly less computational power and memory usage than existing counterparts. Moreover, using various challenging vision datasets we show that the proposed architecture is able to perform one-shot learning – and even use only a small portion of the dataset – while keeping very good accuracy.

Download manuscript.

Bibtex
@inproceedings{HacGriFarArzJez2017,
  author = {Ghouthi Boukli Hacene and Vincent Gripon
and Nicolas Farrugia and Matthieu Arzel and Michel
Jezequel},
  title = {Budget Restricted Incremental Learning with
Pre-Trained Convolutional Neural Networks and Binary
Associative Memories},
  booktitle = {Proceedings of SIPS},
  year = {2017},
  note = {To appear},
}




You are the 531692th visitor

Vincent Gripon's Homepage