1 code implementation • 20 Jul 2023 • Rajiv Movva, Sidhika Balachandar, Kenny Peng, Gabriel Agostini, Nikhil Garg, Emma Pierson
Large language models (LLMs) are dramatically influencing AI research, spurring discussions on what has changed so far and how to shape the field's future.
1 code implementation • 18 Apr 2023 • Rajiv Movva, Divya Shanmugam, Kaihua Hou, Priya Pathak, John Guttag, Nikhil Garg, Emma Pierson
Across outcomes and metrics, we show that the risk scores exhibit significant granular performance disparities within coarse race groups.
no code implementations • COLING 2022 • Rajiv Movva, Jinhao Lei, Shayne Longpre, Ajay Gupta, Chris DuBois
Our work quantitatively demonstrates that combining compression methods can synergistically reduce model size, and that practitioners should prioritize (1) quantization, (2) knowledge distillation, and (3) pruning to maximize accuracy vs. model size tradeoffs.
no code implementations • 30 Apr 2021 • Rajiv Movva, Jonathan Frankle, Michael Carbin
Magnitude pruning is a common, effective technique to identify sparse subnetworks at little cost to accuracy.
no code implementations • EMNLP (BlackboxNLP) 2020 • Rajiv Movva, Jason Y. Zhao
Recent work on the lottery ticket hypothesis has produced highly sparse Transformers for NMT while maintaining BLEU.