no code implementations • bioRxiv 2022 • Zeming Lin, Halil Akin, Roshan Rao, Brian Hie, Zhongkai Zhu, Wenting Lu, Allan dos Santos Costa, Maryam Fazel-Zarandi, Tom Sercu, Sal Candido, Alexander Rives
We find that as models are scaled they learn information enabling the prediction of the three-dimensional structure of a protein at the resolution of individual atoms.
1 code implementation • NeurIPS 2021 • Joshua Meier, Roshan Rao, Robert Verkuil, Jason Liu, Tom Sercu, Alex Rives
Modeling the effect of sequence variation on function is a fundamental problem for understanding and designing proteins.
1 code implementation • 13 Feb 2021 • Roshan Rao, Jason Liu, Robert Verkuil, Joshua Meier, John F. Canny, Pieter Abbeel, Tom Sercu, Alexander Rives
Unsupervised protein language models trained across millions of diverse sequences learn structure and function of proteins.
no code implementations • ICLR 2021 • Roshan Rao, Joshua Meier, Tom Sercu, Sergey Ovchinnikov, Alexander Rives
Unsupervised contact prediction is central to uncovering physical, structural, and functional constraints for protein structure determination and design.
no code implementations • 1 Jan 2021 • Tom Sercu, Robert Verkuil, Joshua Meier, Brandon Amos, Zeming Lin, Caroline Chen, Jason Liu, Yann Lecun, Alexander Rives
We propose the Neural Potts Model objective as an amortized optimization problem.
1 code implementation • Proceedings of the National Academy of Sciences 2020 • Alexander Rives, Joshua Meier, Tom Sercu, Siddharth Goyal, Zeming Lin, Demi Guo, Myle Ott, C. Lawrence Zitnick, Jerry Ma, Rob Fergus
In the field of artificial intelligence, a combination of scale in data and model capacity enabled by unsupervised learning has led to major advances in representation learning and statistical generation.
2 code implementations • 22 May 2020 • Payel Das, Tom Sercu, Kahini Wadhawan, Inkit Padhi, Sebastian Gehrmann, Flaviu Cipcigan, Vijil Chenthamarakshan, Hendrik Strobelt, Cicero dos Santos, Pin-Yu Chen, Yi Yan Yang, Jeremy Tan, James Hedrick, Jason Crain, Aleksandra Mojsilovic
De novo therapeutic design is challenged by a vast chemical repertoire and multiple constraints, e. g., high broad-spectrum potency and low toxicity.
1 code implementation • NeurIPS 2019 • Youssef Mroueh, Tom Sercu, Mattia Rigotti, Inkit Padhi, Cicero dos Santos
In the kernel version we show that SIC can be cast as a convex optimization problem by introducing auxiliary variables that play an important role in feature selection as they are normalized feature importance scores.
no code implementations • 29 Jul 2019 • Tom Sercu, Neil Mallinar
We introduce Multi-Frame Cross-Entropy training (MFCE) for convolutional neural network acoustic models.
no code implementations • ICLR Workshop DeepGenStruct 2019 • Tom Sercu, Sebastian Gehrmann, Hendrik Strobelt, Payel Das, Inkit Padhi, Cicero dos Santos, Kahini Wadhawan, Vijil Chenthamarakshan
We present the pipeline in an interactive visual tool to enable the exploration of the metrics, analysis of the learned latent space, and selection of the best model for a given task.
no code implementations • ICLR Workshop DeepGenStruct 2019 • Pierre Dognin, Igor Melnyk, Youssef Mroueh, Jarret Ross, Tom Sercu
In this paper we study image captioning as a conditional GAN training, proposing both a context-aware LSTM captioner and co-attentive discriminator, which enforces semantic alignment between images and captions.
1 code implementation • 13 Feb 2019 • Pierre Dognin, Igor Melnyk, Youssef Mroueh, Jerret Ross, Cicero dos Santos, Tom Sercu
In this paper we propose to perform model ensembling in a multiclass or a multilabel learning setting using Wasserstein (W.) barycenters.
no code implementations • 17 Oct 2018 • Payel Das, Kahini Wadhawan, Oscar Chang, Tom Sercu, Cicero dos Santos, Matthew Riemer, Vijil Chenthamarakshan, Inkit Padhi, Aleksandra Mojsilovic
Our model learns a rich latent space of the biological peptide context by taking advantage of abundant, unlabeled peptide sequences.
3 code implementations • ICLR 2019 • Chun-Fu Chen, Quanfu Fan, Neil Mallinar, Tom Sercu, Rogerio Feris
The proposed approach demonstrates improvement of model efficiency and performance on both object recognition and speech recognition tasks, using popular architectures including ResNet and ResNeXt.
no code implementations • 30 May 2018 • Youssef Mroueh, Tom Sercu, Anant Raj
We study a simplification of GAN training: the problem of transporting particles from a source to a target distribution.
no code implementations • 30 Apr 2018 • Pierre L. Dognin, Igor Melnyk, Youssef Mroueh, Jarret Ross, Tom Sercu
When evaluated on OOC and MS-COCO benchmarks, we show that SCST-based training has a strong performance in both semantic score and human evaluation, promising to be a valuable new approach for efficient discrete GAN training.
no code implementations • 7 Dec 2017 • Tom Sercu, Youssef Mroueh
We present an empirical investigation of a recent class of Generative Adversarial Networks (GANs) using Integral Probability Metrics (IPM) and their performance for semi-supervised learning.
2 code implementations • ICLR 2018 • Youssef Mroueh, Chun-Liang Li, Tom Sercu, Anant Raj, Yu Cheng
We show that the Sobolev IPM compares two distributions in high dimensions based on weighted conditional Cumulative Distribution Functions (CDF) of each coordinate on a leave one out basis.
2 code implementations • NeurIPS 2017 • Youssef Mroueh, Tom Sercu
In this paper we introduce Fisher GAN which fits within the Integral Probability Metrics (IPM) framework for training GANs.
no code implementations • 6 Mar 2017 • George Saon, Gakuto Kurata, Tom Sercu, Kartik Audhkhasi, Samuel Thomas, Dimitrios Dimitriadis, Xiaodong Cui, Bhuvana Ramabhadran, Michael Picheny, Lynn-Li Lim, Bergul Roomi, Phil Hall
This then raises two issues - what IS human performance, and how far down can we still drive speech recognition error rates?
Ranked #3 on Speech Recognition on Switchboard + Hub500
no code implementations • ICML 2017 • Youssef Mroueh, Tom Sercu, Vaibhava Goel
We introduce new families of Integral Probability Metrics (IPM) for training Generative Adversarial Networks (GAN).
no code implementations • 28 Nov 2016 • Tom Sercu, Vaibhava Goel
We show that dense prediction view of framewise classification offers several advantages and insights, including computational efficiency and the ability to apply batch normalization.
no code implementations • 27 Apr 2016 • George Saon, Tom Sercu, Steven Rennie, Hong-Kwang J. Kuo
We describe a collection of acoustic and language modeling techniques that lowered the word error rate of our English conversational telephone LVCSR system to a record 6. 6% on the Switchboard subset of the Hub5 2000 evaluation testset.
Ranked #5 on Speech Recognition on swb_hub_500 WER fullSWBCH
no code implementations • 6 Apr 2016 • Tom Sercu, Vaibhava Goel
We demonstrate the performance of our models both on larger scale data than before, and after sequence training.
no code implementations • 29 Sep 2015 • Tom Sercu, Christian Puhrsch, Brian Kingsbury, Yann Lecun
However, CNNs in LVCSR have not kept pace with recent advances in other domains where deeper neural networks provide superior performance.
Ranked #17 on Speech Recognition on Switchboard + Hub500