no code implementations • 4 Jun 2024 • Nathan Ng, Roger Grosse, Marzyeh Ghassemi
Estimating the uncertainty of a model's prediction on a test point is a crucial part of ensuring reliability and calibration under distribution shifts.
1 code implementation • 13 Feb 2024 • Kyle O'Brien, Nathan Ng, Isha Puri, Jorge Mendez, Hamid Palangi, Yoon Kim, Marzyeh Ghassemi, Thomas Hartvigsen
Machine learning models for text classification often excel on in-distribution (ID) data but struggle with unseen out-of-distribution (OOD) inputs.
no code implementations • 4 Sep 2023 • Nathan Ng, Ji Won Park, Jae Hyeon Lee, Ryan Lewis Kelly, Stephen Ra, Kyunghyun Cho
This set embedding represents the "average" of the subreads and can be decoded into a prediction of the clean sequence.
2 code implementations • 12 Sep 2022 • Juhan Bae, Nathan Ng, Alston Lo, Marzyeh Ghassemi, Roger Grosse
Influence functions efficiently estimate the effect of removing a single training data point on a model's learned parameters.
no code implementations • 5 Jul 2022 • Nathan Ng, Neha Hulkund, Kyunghyun Cho, Marzyeh Ghassemi
Developing and deploying machine learning models safely depends on the ability to characterize and compare their abilities to generalize to new environments.
no code implementations • 12 Jan 2021 • Nathan Ng, Sebastian Wenderoth, Rajagopala Reddy Seelam, Eran Rabani, Hans-Dieter Meyer, Michael Thoss, Michael Kolodrubetz
At longer time scales, we see slow growth of the entanglement, which may arise from dephasing mechanisms in the localized system or long-range interactions mediated by the central degree of freedom.
Disordered Systems and Neural Networks
no code implementations • 30 Oct 2020 • Nathan Ng, Marzyeh Ghassemi, Narendran Thangarajan, Jiacheng Pan, Qi Guo
In ablations on DBDC4 data from 2019, our semi-supervised learning methods improve the performance of a baseline BERT model by 2\% accuracy.
1 code implementation • EMNLP 2020 • Nathan Ng, Kyunghyun Cho, Marzyeh Ghassemi
Models that perform well on a training domain often fail to generalize to out-of-domain (OOD) examples.
1 code implementation • IJCNLP 2019 • Kyra Yee, Nathan Ng, Yann N. Dauphin, Michael Auli
Previous work on neural noisy channel modeling relied on latent variable models that incrementally process the source and target sentence.
5 code implementations • WS 2019 • Nathan Ng, Kyra Yee, Alexei Baevski, Myle Ott, Michael Auli, Sergey Edunov
This paper describes Facebook FAIR's submission to the WMT19 shared news translation task.
Ranked #1 on Machine Translation on WMT2019 English-German
no code implementations • 9 Apr 2019 • Tingfung Lau, Nathan Ng, Julian Gingold, Nina Desai, Julian McAuley, Zachary C. Lipton
First, noting that in each image the embryo occupies a small subregion, we jointly train a region proposal network with the downstream classifier to isolate the embryo.
6 code implementations • NAACL 2019 • Myle Ott, Sergey Edunov, Alexei Baevski, Angela Fan, Sam Gross, Nathan Ng, David Grangier, Michael Auli
fairseq is an open-source sequence modeling toolkit that allows researchers and developers to train custom models for translation, summarization, language modeling, and other text generation tasks.
no code implementations • 17 Feb 2017 • Nathan Ng, Rodney A Gabriel, Julian McAuley, Charles Elkan, Zachary C. Lipton
Scheduling surgeries is a challenging task due to the fundamental uncertainty of the clinical environment, as well as the risks and costs associated with under- and over-booking.