no code implementations • 19 Jan 2021 • Ye Ouyang, Lilei Wang, Aidong Yang, Maulik Shah, David Belanger, Tongqing Gao, Leping Wei, Yaqin Zhang
It has been an exciting journey since the mobile communications and artificial intelligence were conceived 37 years and 64 years ago.
no code implementations • 31 Oct 2020 • Amir Shanehsazzadeh, David Belanger, David Dohan
In this paper, we show that CNN models trained solely using supervised learning both compete with and sometimes outperform the best models from TAPE that leverage expensive pretraining on large protein datasets.
1 code implementation • 15 Oct 2020 • Amir Shanehsazzadeh, David Belanger, David Dohan
We consider transformer (BERT) protein language models that are pretrained on the TrEMBL data set and learn fixed-length embeddings on top of them with contextual lenses.
7 code implementations • ICLR 2021 • Krzysztof Choromanski, Valerii Likhosherstov, David Dohan, Xingyou Song, Andreea Gane, Tamas Sarlos, Peter Hawkins, Jared Davis, Afroz Mohiuddin, Lukasz Kaiser, David Belanger, Lucy Colwell, Adrian Weller
We introduce Performers, Transformer architectures which can estimate regular (softmax) full-rank-attention Transformers with provable accuracy, but using only linear (as opposed to quadratic) space and time complexity, without relying on any priors such as sparsity or low-rankness.
Ranked #7 on Offline RL on D4RL
1 code implementation • 5 Jun 2020 • Krzysztof Choromanski, Valerii Likhosherstov, David Dohan, Xingyou Song, Andreea Gane, Tamas Sarlos, Peter Hawkins, Jared Davis, David Belanger, Lucy Colwell, Adrian Weller
In response, solutions that exploit the structure and sparsity of the learned attention matrix have blossomed.
no code implementations • ICML 2020 • Christof Angermueller, David Belanger, Andreea Gane, Zelda Mariet, David Dohan, Kevin Murphy, Lucy Colwell, D. Sculley
The cost and latency of wet-lab experiments requires methods that find good sequences in few experimental rounds of large batches of sequences--a setting that off-the-shelf black-box optimization methods are ill-equipped to handle.
no code implementations • ICLR 2020 • Christof Angermueller, David Dohan, David Belanger, Ramya Deshpande, Kevin Murphy, Lucy Colwell
In response, we propose using reinforcement learning (RL) based on proximal-policy optimization (PPO) for biological sequence design.
Model-based Reinforcement Learning reinforcement-learning +2
no code implementations • ICCV 2019 • Piotr Teterwak, Aaron Sarna, Dilip Krishnan, Aaron Maschinot, David Belanger, Ce Liu, William T. Freeman
Image extension models have broad applications in image editing, computational photography and computer graphics.
Ranked #2 on Uncropping on Places2 val
no code implementations • 21 Nov 2018 • Jennifer N. Wei, David Belanger, Ryan P. Adams, D. Sculley
When confronted with a substance of unknown identity, researchers often perform mass spectrometry on the sample and compare the observed spectrum to a library of previously-collected spectra to identify the molecule.
2 code implementations • ICLR 2018 • Gonzalo Mena, David Belanger, Scott Linderman, Jasper Snoek
Permutations and matchings are core building blocks in a variety of latent variable models, as they allow us to align, canonicalize, and sort data.
no code implementations • 2 Aug 2017 • Dung Thai, Shikhar Murty, Trapit Bansal, Luke Vilnis, David Belanger, Andrew McCallum
In textual information extraction and other sequence labeling tasks it is now common to use recurrent neural networks (such as LSTM) to form rich embedded representations of long-term input co-occurrence patterns.
no code implementations • ICML 2017 • David Belanger, Bishan Yang, Andrew McCallum
Structured Prediction Energy Networks (SPENs) are a simple, yet expressive family of structured prediction models (Belanger and McCallum, 2016).
4 code implementations • EMNLP 2017 • Emma Strubell, Patrick Verga, David Belanger, Andrew McCallum
Today when many practitioners run basic NLP on the entire web and large-volume traffic, faster methods are paramount to saving time and energy costs.
Ranked #25 on Named Entity Recognition (NER) on Ontonotes v5 (English)
1 code implementation • CVPR 2017 • Forrester Cole, David Belanger, Dilip Krishnan, Aaron Sarna, Inbar Mosseri, William T. Freeman
We present a method for synthesizing a frontal, neutral-expression image of a person's face given an input face photograph.
no code implementations • 7 Sep 2016 • Trapit Bansal, David Belanger, Andrew McCallum
In a variety of application domains the content to be recommended to users is associated with text.
2 code implementations • EACL 2017 • Rajarshi Das, Arvind Neelakantan, David Belanger, Andrew McCallum
Our goal is to combine the rich multistep inference of symbolic logical reasoning with the generalization capabilities of neural networks.
1 code implementation • NAACL 2016 • Patrick Verga, David Belanger, Emma Strubell, Benjamin Roth, Andrew McCallum
In response, this paper introduces significant further improvements to the coverage and flexibility of universal schema relation extraction: predictions for entities unseen in training and multilingual transfer learning to domains with no annotation.
no code implementations • 19 Nov 2015 • David Belanger, Andrew McCallum
This deep architecture captures dependencies between labels that would lead to intractable graphical models, and performs structure learning by automatically learning discriminative features of the structured output.
no code implementations • 4 Mar 2015 • Kui Tang, Nicholas Ruozzi, David Belanger, Tony Jebara
Many machine learning tasks can be formulated in terms of predicting structured outputs.
no code implementations • 4 Mar 2015 • Luke Vilnis, David Belanger, Daniel Sheldon, Andrew McCallum
Many inference problems in structured prediction are naturally solved by augmenting a tractable dependency structure with complex, non-local auxiliary objectives.
no code implementations • 13 Feb 2015 • David Belanger, Sham Kakade
Finally, the Kalman filter updates can be seen as a linear recurrent neural network.
no code implementations • ACL 2014 • Sam Anzaroot, Alexandre Passos, David Belanger, Andrew McCallum
Accurately segmenting a citation string into fields for authors, titles, etc.
no code implementations • NeurIPS 2012 • David Belanger, Alexandre Passos, Sebastian Riedel, Andrew McCallum
Linear chains and trees are basic building blocks in many applications of graphical models.