no code implementations • 11 May 2022 • Dimitris Stripelis, Umang Gupta, Hamza Saleem, Nikhil Dhinagar, Tanmay Ghai, Rafael Sanchez, Chrysovalantis Anastasiou, Armaghan Asghar, Greg Ver Steeg, Srivatsan Ravi, Muhammad Naveed, Paul M. Thompson, Jose Luis Ambite
Specifically, we investigate training neural models to classify Alzheimer's disease, and estimate Brain Age, from magnetic resonance imaging datasets distributed across multiple sites, including heterogeneous environments where sites have different amounts of data, statistical distributions, and computational capabilities.
no code implementations • 26 Apr 2022 • Dimitris Stripelis, Umang Gupta, Greg Ver Steeg, Jose Luis Ambite
Second, the models are incrementally constrained to a smaller set of parameters, which facilitates alignment/merging of the local models and improved learning performance at high sparsification rates.
no code implementations • Findings (ACL) 2022 • Umang Gupta, Jwala Dhamala, Varun Kumar, Apurv Verma, Yada Pruksachatkun, Satyapriya Krishna, Rahul Gupta, Kai-Wei Chang, Greg Ver Steeg, Aram Galstyan
Language models excel at generating coherent text, and model compression techniques such as knowledge distillation have enabled their use in resource-constrained settings.
no code implementations • 8 Sep 2021 • Ninareh Mehrabi, Umang Gupta, Fred Morstatter, Greg Ver Steeg, Aram Galstyan
The widespread use of Artificial Intelligence (AI) in consequential domains, such as healthcare and parole decision-making systems, has drawn intense scrutiny on the fairness of these methods.
no code implementations • 7 Aug 2021 • Dimitris Stripelis, Hamza Saleem, Tanmay Ghai, Nikhil Dhinagar, Umang Gupta, Chrysovalantis Anastasiou, Greg Ver Steeg, Srivatsan Ravi, Muhammad Naveed, Paul M. Thompson, Jose Luis Ambite
Federated learning (FL) enables distributed computation of machine learning models over various disparate, remote data sources, without requiring to transfer any individual data to a centralized location.
no code implementations • 6 May 2021 • Umang Gupta, Dimitris Stripelis, Pradeep K. Lam, Paul M. Thompson, José Luis Ambite, Greg Ver Steeg
In particular, we show that it is possible to infer if a sample was used to train the model given only access to the model prediction (black-box) or access to the model itself (white-box) and some leaked samples from the training data distribution.
1 code implementation • 8 Feb 2021 • Umang Gupta, Pradeep K. Lam, Greg Ver Steeg, Paul M. Thompson
Deep Learning for neuroimaging data is a promising but challenging direction.
2 code implementations • 11 Jan 2021 • Umang Gupta, Aaron M Ferber, Bistra Dilkina, Greg Ver Steeg
Controlling bias in training datasets is vital for ensuring equal treatment, or parity, between different groups in downstream applications.
no code implementations • ICLR 2018 • Nitin Kamra, Umang Gupta, Yan Liu
This phenomenon called catastrophic forgetting is a fundamental challenge to overcome before neural networks can learn continually from incoming data.
no code implementations • 21 Jul 2017 • Umang Gupta, Ankush Chatterjee, Radhakrishnan Srikanth, Puneet Agrawal
In this paper, we propose a novel approach to detect emotions like happy, sad or angry in textual conversations using an LSTM based Deep Learning model.