no code implementations • 6 Nov 2024 • Marlon Tobaben, Mohamed Ali Souibgui, Rubèn Tito, Khanh Nguyen, Raouf Kerkouche, Kangsoo Jung, Joonas Jälkö, Lei Kang, Andrey Barsky, Vincent Poulain D'Andecy, Aurélie Joseph, Aashiq Muhamed, Kevin Kuo, Virginia Smith, Yusuke Yamasaki, Takumi Fukami, Kenta Niwa, Iifan Tyou, Hiro Ishii, Rio Yokota, Ragul N, Rintu Kutum, Josep Llados, Ernest Valveny, Antti Honkela, Mario Fritz, Dimosthenis Karatzas
The Privacy Preserving Federated Learning Document VQA (PFL-DocVQA) competition challenged the community to develop provably private and communication-efficient solutions in a federated setting for a real-life use case: invoice processing.
no code implementations • 1 Nov 2024 • Aashiq Muhamed, Mona Diab, Virginia Smith
Understanding and mitigating the potential risks associated with foundation models (FMs) hinges on developing effective interpretability methods.
no code implementations • 3 Oct 2024 • Pratiksha Thaker, Shengyuan Hu, Neil Kale, Yash Maurya, Zhiwei Steven Wu, Virginia Smith
Unlearning methods have the potential to improve the privacy and safety of large language models (LLMs) by removing sensitive or harmful information post hoc.
no code implementations • 2 Jul 2024 • Steven Kolawole, Don Dennis, Ameet Talwalkar, Virginia Smith
Adaptive inference schemes reduce the cost of machine learning inference by assigning smaller models to easier examples, attempting to avoid invocation of larger models when possible.
1 code implementation • 25 Jun 2024 • Aashiq Muhamed, Oscar Li, David Woodruff, Mona Diab, Virginia Smith
Large language model (LLM) training and finetuning are often bottlenecked by limited GPU memory.
1 code implementation • 20 Jun 2024 • Amrith Setlur, Saurabh Garg, Xinyang Geng, Naman Garg, Virginia Smith, Aviral Kumar
With this per-step scheme, we are able to attain consistent gains over only positive data, attaining performance similar to amplifying the amount of synthetic data by $\mathbf{8 \times}$.
1 code implementation • 19 Jun 2024 • Shengyuan Hu, Yiwei Fu, Zhiwei Steven Wu, Virginia Smith
Machine unlearning is a promising approach to mitigate undesirable memorization of training data in LLMs.
1 code implementation • 7 Jun 2024 • Kevin Kuo, Arian Raje, Kousik Rajesh, Virginia Smith
Prior work that has studied LoRA in the context of federated learning has focused on improving LoRA's robustness to heterogeneity and privacy.
no code implementations • 7 Mar 2024 • Shengyuan Hu, Saeed Mahloujifar, Virginia Smith, Kamalika Chaudhuri, Chuan Guo
Data-dependent privacy accounting frameworks such as per-instance differential privacy (pDP) and Fisher information loss (FIL) confer fine-grained privacy guarantees for individuals in a fixed training dataset.
no code implementations • 6 Mar 2024 • Ziyue Li, Tian Li, Virginia Smith, Jeff Bilmes, Tianyi Zhou
Optimizing the performance of many objectives (instantiated by tasks or clients) jointly with a few Pareto stationary solutions (models) is critical in machine learning.
1 code implementation • 5 Mar 2024 • Pratiksha Thaker, Yash Maurya, Shengyuan Hu, Zhiwei Steven Wu, Virginia Smith
Recent work has demonstrated that finetuning is a promising approach to 'unlearn' concepts from large language models.
1 code implementation • 25 Feb 2024 • Qi Pang, Shengyuan Hu, Wenting Zheng, Virginia Smith
Advances in generative models have made it possible for AI-generated text, code, and images to mirror human-generated content in many applications.
1 code implementation • 8 Feb 2024 • Lucio Dery, Steven Kolawole, Jean-François Kagy, Virginia Smith, Graham Neubig, Ameet Talwalkar
Given the generational gap in available hardware between lay practitioners and the most endowed institutions, LLMs are becoming increasingly inaccessible as they grow in size.
no code implementations • 24 Dec 2023 • Pratiksha Thaker, Amrith Setlur, Zhiwei Steven Wu, Virginia Smith
Public pretraining is a promising approach to improve differentially private model training.
no code implementations • NeurIPS 2023 • Saurabh Garg, Amrith Setlur, Zachary Chase Lipton, Sivaraman Balakrishnan, Virginia Smith, aditi raghunathan
Self-training and contrastive learning have emerged as leading techniques for incorporating unlabeled data, both under distribution shift (unsupervised domain adaptation) and when it is absent (semi-supervised learning).
1 code implementation • NeurIPS 2023 • Oscar Li, James Harrison, Jascha Sohl-Dickstein, Virginia Smith, Luke Metz
Unrolled computation graphs are prevalent throughout machine learning but present challenges to automatic differentiation (AD) gradient estimation methods when their loss functions exhibit extreme local sensitivtiy, discontinuity, or blackbox characteristics.
no code implementations • 16 Feb 2023 • Shengyuan Hu, Dung Daniel Ngo, Shuran Zheng, Virginia Smith, Zhiwei Steven Wu
Federated Learning (FL) aims to foster collaboration among a population of clients to improve the accuracy of machine learning without directly sharing local data.
1 code implementation • 6 Feb 2023 • Amrith Setlur, Don Dennis, Benjamin Eysenbach, aditi raghunathan, Chelsea Finn, Virginia Smith, Sergey Levine
Some robust training algorithms (e. g., Group DRO) specialize to group shifts and require group information on all training points.
1 code implementation • 17 Dec 2022 • Kevin Kuo, Pratiksha Thaker, Mikhail Khodak, John Nguyen, Daniel Jiang, Ameet Talwalkar, Virginia Smith
In this work, we perform the first systematic study on the effect of noisy evaluation in federated hyperparameter tuning.
1 code implementation • 1 Dec 2022 • Tian Li, Manzil Zaheer, Ken Ziyu Liu, Sashank J. Reddi, H. Brendan McMahan, Virginia Smith
Privacy noise may negate the benefits of using adaptive optimizers in differentially private model training.
1 code implementation • 21 Nov 2022 • Michael Kuchnik, Virginia Smith, George Amvrosiadis
Although large language models (LLMs) have been touted for their ability to generate natural-sounding text, there are growing concerns around possible negative effects of LLMs such as data memorization, bias, and inappropriate language.
2 code implementations • 18 Jun 2022 • Shanshan Wu, Tian Li, Zachary Charles, Yu Xiao, Ziyu Liu, Zheng Xu, Virginia Smith
To better answer these questions, we propose Motley, a benchmark for personalized federated learning.
1 code implementation • 16 Jun 2022 • Ziyu Liu, Shengyuan Hu, Zhiwei Steven Wu, Virginia Smith
While the application of differential privacy (DP) has been well-studied in cross-device federated learning (FL), there is a lack of work considering DP and its implications for cross-silo FL, a setting characterized by a limited number of clients each containing many data subjects.
no code implementations • 3 Jun 2022 • Amrith Setlur, Benjamin Eysenbach, Virginia Smith, Sergey Levine
Supervised learning methods trained with maximum likelihood objectives often overfit on training data.
no code implementations • 30 May 2022 • Yae Jee Cho, Divyansh Jhunjhunwala, Tian Li, Virginia Smith, Gauri Joshi
We provide convergence guarantees for MaxFL and show that MaxFL achieves a $22$-$40\%$ and $18$-$50\%$ test accuracy improvement for the training clients and unseen clients respectively, compared to a wide range of FL modeling approaches, including those that tackle data heterogeneity, aim to incentivize clients, and learn personalized or fair models.
no code implementations • 18 Mar 2022 • Shengyuan Hu, Zhiwei Steven Wu, Virginia Smith
In particular, we explore and extend the notion of Bounded Group Loss as a theoretically-grounded approach for group fairness.
1 code implementation • 12 Feb 2022 • Tian Li, Manzil Zaheer, Sashank J. Reddi, Virginia Smith
Adaptive optimization methods have become the default solvers for many machine learning tasks.
2 code implementations • 7 Nov 2021 • Michael Kuchnik, Ana Klimovic, Jiri Simsa, Virginia Smith, George Amvrosiadis
Our analysis of over two million ML jobs in Google datacenters reveals that a significant fraction of model training jobs could benefit from faster input data pipelines.
no code implementations • ICLR 2022 • Ravikumar Balakrishnan, Tian Li, Tianyi Zhou, Nageen Himayat, Virginia Smith, Jeff Bilmes
In every communication round of federated learning, a random subset of clients communicate their model updates back to the server which then aggregates them all.
1 code implementation • 13 Sep 2021 • Tian Li, Ahmad Beirami, Maziar Sanjabi, Virginia Smith
Finally, we demonstrate that TERM can be used for a multitude of applications in machine learning, such as enforcing fairness between subgroups, mitigating the effect of outliers, and handling class imbalance.
1 code implementation • 30 Aug 2021 • Shengyuan Hu, Zhiwei Steven Wu, Virginia Smith
Many problems in machine learning rely on multi-task learning (MTL), in which the goal is to solve multiple related machine learning tasks simultaneously.
2 code implementations • 14 Jul 2021 • Jianyu Wang, Zachary Charles, Zheng Xu, Gauri Joshi, H. Brendan McMahan, Blaise Aguera y Arcas, Maruan Al-Shedivat, Galen Andrew, Salman Avestimehr, Katharine Daly, Deepesh Data, Suhas Diggavi, Hubert Eichner, Advait Gadhikar, Zachary Garrett, Antonious M. Girgis, Filip Hanzely, Andrew Hard, Chaoyang He, Samuel Horvath, Zhouyuan Huo, Alex Ingerman, Martin Jaggi, Tara Javidi, Peter Kairouz, Satyen Kale, Sai Praneeth Karimireddy, Jakub Konecny, Sanmi Koyejo, Tian Li, Luyang Liu, Mehryar Mohri, Hang Qi, Sashank J. Reddi, Peter Richtarik, Karan Singhal, Virginia Smith, Mahdi Soltanolkotabi, Weikang Song, Ananda Theertha Suresh, Sebastian U. Stich, Ameet Talwalkar, Hongyi Wang, Blake Woodworth, Shanshan Wu, Felix X. Yu, Honglin Yuan, Manzil Zaheer, Mi Zhang, Tong Zhang, Chunxiang Zheng, Chen Zhu, Wennan Zhu
Federated learning and analytics are a distributed approach for collaboratively learning models (or statistics) from decentralized data, motivated by and designed for privacy protection.
no code implementations • NeurIPS 2021 • Zachary Charles, Zachary Garrett, Zhouyuan Huo, Sergei Shmulyian, Virginia Smith
Our work highlights a number of challenges stemming from the use of larger cohorts.
no code implementations • NeurIPS 2021 • Mikhail Khodak, Renbo Tu, Tian Li, Liam Li, Maria-Florina Balcan, Virginia Smith, Ameet Talwalkar
Tuning hyperparameters is a crucial but arduous part of the machine learning pipeline.
2 code implementations • 1 Mar 2021 • Don Kurian Dennis, Tian Li, Virginia Smith
In this work, we explore the unique challenges -- and opportunities -- of unsupervised federated learning (FL).
1 code implementation • NeurIPS 2021 • Amrith Setlur, Oscar Li, Virginia Smith
We categorize meta-learning evaluation into two settings: $\textit{in-distribution}$ [ID], in which the train and test tasks are sampled $\textit{iid}$ from the same underlying task distribution, and $\textit{out-of-distribution}$ [OOD], in which they are not.
2 code implementations • ICLR 2022 • Oscar Li, Jiankai Sun, Xin Yang, Weihao Gao, Hongyi Zhang, Junyuan Xie, Virginia Smith, Chong Wang
Two-party split learning is a popular technique for learning a model across feature-partitioned data.
4 code implementations • 8 Dec 2020 • Tian Li, Shengyuan Hu, Ahmad Beirami, Virginia Smith
Fairness and robustness are two important concerns for federated learning systems.
no code implementations • 28 Nov 2020 • Amrith Setlur, Oscar Li, Virginia Smith
Meta-learning is a popular framework for learning with limited data in which an algorithm is produced by training over multiple few-shot learning tasks.
2 code implementations • ICLR 2021 • Tian Li, Ahmad Beirami, Maziar Sanjabi, Virginia Smith
Empirical risk minimization (ERM) is typically designed to perform well on the average loss, which can result in estimators that are sensitive to outliers, generalize poorly, or treat subgroups unfairly.
2 code implementations • 7 Jan 2020 • Tian Li, Anit Kumar Sahu, Manzil Zaheer, Maziar Sanjabi, Ameet Talwalkar, Virginia Smith
Federated learning aims to jointly learn statistical models over massively distributed remote devices.
no code implementations • 5 Nov 2019 • Zaoxing Liu, Tian Li, Virginia Smith, Vyas Sekar
Federated learning methods run training tasks directly on user devices and do not share the raw user data with third parties.
no code implementations • 3 Nov 2019 • Tian Li, Zaoxing Liu, Vyas Sekar, Virginia Smith
Many existing works treat these concerns separately.
1 code implementation • 1 Nov 2019 • Michael Kuchnik, George Amvrosiadis, Virginia Smith
Deep learning accelerators efficiently train over vast and growing amounts of data, placing a newfound burden on commodity networks and storage devices.
1 code implementation • 21 Aug 2019 • Tian Li, Anit Kumar Sahu, Ameet Talwalkar, Virginia Smith
Federated learning involves training statistical models over remote devices or siloed data centers, such as mobile phones or hospitals, while keeping data localized.
3 code implementations • ICLR 2020 • Tian Li, Maziar Sanjabi, Ahmad Beirami, Virginia Smith
Federated learning involves training statistical models in massive, heterogeneous networks.
no code implementations • 29 Mar 2019 • Alexander Ratner, Dan Alistarh, Gustavo Alonso, David G. Andersen, Peter Bailis, Sarah Bird, Nicholas Carlini, Bryan Catanzaro, Jennifer Chayes, Eric Chung, Bill Dally, Jeff Dean, Inderjit S. Dhillon, Alexandros Dimakis, Pradeep Dubey, Charles Elkan, Grigori Fursin, Gregory R. Ganger, Lise Getoor, Phillip B. Gibbons, Garth A. Gibson, Joseph E. Gonzalez, Justin Gottschlich, Song Han, Kim Hazelwood, Furong Huang, Martin Jaggi, Kevin Jamieson, Michael. I. Jordan, Gauri Joshi, Rania Khalaf, Jason Knight, Jakub Konečný, Tim Kraska, Arun Kumar, Anastasios Kyrillidis, Aparna Lakshmiratan, Jing Li, Samuel Madden, H. Brendan McMahan, Erik Meijer, Ioannis Mitliagkas, Rajat Monga, Derek Murray, Kunle Olukotun, Dimitris Papailiopoulos, Gennady Pekhimenko, Theodoros Rekatsinas, Afshin Rostamizadeh, Christopher Ré, Christopher De Sa, Hanie Sedghi, Siddhartha Sen, Virginia Smith, Alex Smola, Dawn Song, Evan Sparks, Ion Stoica, Vivienne Sze, Madeleine Udell, Joaquin Vanschoren, Shivaram Venkataraman, Rashmi Vinayak, Markus Weimer, Andrew Gordon Wilson, Eric Xing, Matei Zaharia, Ce Zhang, Ameet Talwalkar
Machine learning (ML) techniques are enjoying rapidly increasing adoption.
no code implementations • 28 Feb 2019 • Neel Guha, Ameet Talwalkar, Virginia Smith
We present one-shot federated learning, where a central server learns a global model over a network of federated devices in a single round of communication.
22 code implementations • 14 Dec 2018 • Tian Li, Anit Kumar Sahu, Manzil Zaheer, Maziar Sanjabi, Ameet Talwalkar, Virginia Smith
Theoretically, we provide convergence guarantees for our framework when learning over data from non-identical distributions (statistical heterogeneity), and while adhering to device-level systems constraints by allowing each participating device to perform a variable amount of work (systems heterogeneity).
7 code implementations • 3 Dec 2018 • Sebastian Caldas, Sai Meher Karthik Duddu, Peter Wu, Tian Li, Jakub Konečný, H. Brendan McMahan, Virginia Smith, Ameet Talwalkar
Modern federated networks, such as those comprised of wearable devices, mobile phones, or autonomous vehicles, generate massive amounts of data each day.
no code implementations • ICLR 2019 • Michael Kuchnik, Virginia Smith
Data augmentation is commonly used to encode invariances in learning methods.
no code implementations • 20 May 2018 • Neel Guha, Virginia Smith
In this work, we present Good-Enough Model Spaces (GEMS), a novel framework for learning a global model by carefully intersecting the sets of "good-enough" models across each node.
no code implementations • 16 Mar 2018 • Tri Dao, Albert Gu, Alexander J. Ratner, Virginia Smith, Christopher De Sa, Christopher Ré
Data augmentation, a technique in which a training set is expanded with class-preserving transformations, is ubiquitous in modern machine learning pipelines.
2 code implementations • NeurIPS 2017 • Virginia Smith, Chao-Kai Chiang, Maziar Sanjabi, Ameet Talwalkar
Federated learning poses new statistical and systems challenges in training machine learning models over distributed networks of devices.
2 code implementations • 7 Nov 2016 • Virginia Smith, Simone Forte, Chenxin Ma, Martin Takac, Michael. I. Jordan, Martin Jaggi
The scale of modern datasets necessitates the development of efficient distributed optimization methods for machine learning.
1 code implementation • 13 Dec 2015 • Chenxin Ma, Jakub Konečný, Martin Jaggi, Virginia Smith, Michael. I. Jordan, Peter Richtárik, Martin Takáč
To this end, we present a framework for distributed optimization that both allows the flexibility of arbitrary solvers to be used on each (single) machine locally, and yet maintains competitive performance against other state-of-the-art special-purpose distributed methods.
2 code implementations • 13 Dec 2015 • Virginia Smith, Simone Forte, Michael. I. Jordan, Martin Jaggi
Despite the importance of sparsity in many large-scale applications, there are few methods for distributed optimization of sparsity-inducing objectives.
1 code implementation • 12 Feb 2015 • Chenxin Ma, Virginia Smith, Martin Jaggi, Michael. I. Jordan, Peter Richtárik, Martin Takáč
Distributed optimization methods for large-scale machine learning suffer from a communication bottleneck.
no code implementations • NeurIPS 2014 • Martin Jaggi, Virginia Smith, Martin Takáč, Jonathan Terhorst, Sanjay Krishnan, Thomas Hofmann, Michael. I. Jordan
Communication remains the most significant bottleneck in the performance of distributed optimization algorithms for large-scale machine learning.
no code implementations • 21 Oct 2013 • Evan R. Sparks, Ameet Talwalkar, Virginia Smith, Jey Kottalam, Xinghao Pan, Joseph Gonzalez, Michael J. Franklin, Michael. I. Jordan, Tim Kraska
MLI is an Application Programming Interface designed to address the challenges of building Machine Learn- ing algorithms in a distributed setting based on data-centric computing.