no code implementations • ICLR 2019 • Yassine Benyahia, Kaicheng Yu, Kamil Bennani-Smires, Martin Jaggi, Anthony Davison, Mathieu Salzmann, Claudiu Musat
We identify a phenomenon, which we refer to as multi-model forgetting, that occurs when sequentially training multiple deep networks with partially-shared parameters; the performance of previously-trained models degrades as one optimizes a subsequent one, due to the overwriting of shared parameters.
no code implementations • 16 Jan 2018 • Kamil Bennani-Smires, Claudiu Musat, Andreea Hossmann, Michael Baeriswyl
The dramatic success of deep neural networks across multiple application areas often relies on experts painstakingly designing a network architecture specific to each task.
3 code implementations • CONLL 2018 • Kamil Bennani-Smires, Claudiu Musat, Andreea Hossmann, Michael Baeriswyl, Martin Jaggi
EmbedRank achieves higher F-scores than graph-based state of the art systems on standard datasets and is suitable for real-time processing of large amounts of Web data.