1 code implementation • 16 Oct 2024 • Genta Indra Winata, Frederikus Hudi, Patrick Amadeus Irawan, David Anugraha, Rifki Afina Putri, Yutong Wang, Adam Nohejl, Ubaidillah Ariq Prathama, Nedjma Ousidhoum, Afifa Amriani, Anar Rzayev, Anirban Das, Ashmari Pramodya, Aulia Adila, Bryan Wilie, Candy Olivia Mawalim, Ching Lam Cheng, Daud Abolade, Emmanuele Chersoni, Enrico Santus, Fariz Ikhwantri, Garry Kuwanto, Hanyang Zhao, Haryo Akbarianto Wibowo, Holy Lovenia, Jan Christian Blaise Cruz, Jan Wira Gotama Putra, Junho Myung, Lucky Susanto, Maria Angelica Riera Machin, Marina Zhukova, Michael Anugraha, Muhammad Farid Adilazuarda, Natasha Santosa, Peerat Limkonchotiwat, Raj Dabre, Rio Alexander Audino, Samuel Cahyawijaya, Shi-Xiong Zhang, Stephanie Yulia Salim, Yi Zhou, Yinxuan Gui, David Ifeoluwa Adelani, En-Shiun Annie Lee, Shogo Okada, Ayu Purwarianti, Alham Fikri Aji, Taro Watanabe, Derry Tanti Wijaya, Alice Oh, Chong-Wah Ngo
This benchmark includes a visual question answering (VQA) dataset with text-image pairs across 30 languages and dialects, spanning 9 language families and featuring over 1 million data points, making it the largest multicultural VQA benchmark to date.
no code implementations • 5 Oct 2024 • Hanyang Zhao, Genta Indra Winata, Anirban Das, Shi-Xiong Zhang, David D. Yao, Wenpin Tang, Sambit Sahu
Recently, numerous preference optimization algorithms have been introduced as extensions to the Direct Preference Optimization (DPO) family.
no code implementations • 19 Sep 2024 • Akshaj Kumar Veldanda, Shi-Xiong Zhang, Anirban Das, Supriyo Chakraborty, Stephen Rawls, Sambit Sahu, Milind Naphade
Large language models (LLMs) have revolutionized various domains, yet their utility comes with significant challenges related to outdated or problematic knowledge embedded during pretraining.
no code implementations • 17 Sep 2024 • Genta Indra Winata, Hanyang Zhao, Anirban Das, Wenpin Tang, David D. Yao, Shi-Xiong Zhang, Sambit Sahu
Preference tuning is a crucial process for aligning deep generative models with human preferences.
no code implementations • 16 Jun 2022 • Timothy Castiglia, Anirban Das, Shiqiang Wang, Stacy Patterson
Our work provides the first theoretical analysis of the effect message compression has on distributed training over vertically partitioned data.
no code implementations • 19 Aug 2021 • Anirban Das, Timothy Castiglia, Shiqiang Wang, Stacy Patterson
Each silo contains a hub and a set of clients, with the silo's vertical data shard partitioned horizontally across its clients.
no code implementations • 6 Feb 2021 • Anirban Das, Stacy Patterson
Each silo contains a hub and a set of clients, with the silo's vertical data shard partitioned horizontally across its clients.
no code implementations • ICLR 2021 • Timothy Castiglia, Anirban Das, Stacy Patterson
We propose Multi-Level Local SGD, a distributed stochastic gradient method for learning a smooth, non-convex objective in a multi-level communication network with heterogeneous workers.
1 code implementation • 27 Jul 2020 • Timothy Castiglia, Anirban Das, Stacy Patterson
In our algorithm, sub-networks execute a distributed SGD algorithm, using a hub-and-spoke paradigm, and the hubs periodically average their models with neighboring hubs.
no code implementations • 11 Nov 2019 • Anirban Das, Thomas Brunschwiler
Federated Learning enables training of a general model through edge devices without sending raw data to the cloud.
1 code implementation • 14 Nov 2018 • Anirban Das, Stacy Patterson, Mike P. Wittie
The emerging trend of edge computing has led several cloud providers to release their own platforms for performing computation at the 'edge' of the network.
Networking and Internet Architecture Distributed, Parallel, and Cluster Computing