no code implementations • 5 Dec 2024 • Dennis Wei, Inkit Padhi, Soumya Ghosh, Amit Dhurandhar, Karthikeyan Natesan Ramamurthy, Maria Chang
To serve as a gold standard for TDA in this "final-model-only" setting, we propose further training, with appropriate adjustment and averaging, to measure the sensitivity of the given model to training instances.
no code implementations • 8 Jul 2024 • Manish Nagireddy, Inkit Padhi, Soumya Ghosh, Prasanna Sattigeri
Motivated by findings from developing a detector for social bias, we adopt the notion of a use-mention distinction - which we identified as the primary source of under-performance in the preliminary versions of our social bias detector.
no code implementations • 1 Jun 2024 • Tejaswini Pedapati, Amit Dhurandhar, Soumya Ghosh, Soham Dan, Prasanna Sattigeri
Estimating uncertainty or confidence in the responses of a model can be significant in evaluating trust not only in the responses, but also in the model as a whole.
no code implementations • 27 May 2024 • Runqian Wang, Soumya Ghosh, David Cox, Diego Antognini, Aude Oliva, Rogerio Feris, Leonid Karlinsky
Our approach relies on synthetic data to transfer LoRA modules.
no code implementations • 21 Mar 2024 • Lucas Monteiro Paes, Dennis Wei, Hyo Jin Do, Hendrik Strobelt, Ronny Luss, Amit Dhurandhar, Manish Nagireddy, Karthikeyan Natesan Ramamurthy, Prasanna Sattigeri, Werner Geyer, Soumya Ghosh
To address the challenges of text as output and long text inputs, we propose a general framework called MExGen that can be instantiated with different attribution algorithms.
no code implementations • 9 Mar 2024 • Swapnaja Achintalwar, Adriana Alvarado Garcia, Ateret Anaby-Tavor, Ioana Baldini, Sara E. Berger, Bishwaranjan Bhattacharjee, Djallel Bouneffouf, Subhajit Chaudhury, Pin-Yu Chen, Lamogha Chiazor, Elizabeth M. Daly, Kirushikesh DB, Rogério Abreu de Paula, Pierre Dognin, Eitan Farchi, Soumya Ghosh, Michael Hind, Raya Horesh, George Kour, Ja Young Lee, Nishtha Madaan, Sameep Mehta, Erik Miehling, Keerthiram Murugesan, Manish Nagireddy, Inkit Padhi, David Piorkowski, Ambrish Rawat, Orna Raz, Prasanna Sattigeri, Hendrik Strobelt, Sarathkrishna Swaminathan, Christoph Tillmann, Aashka Trivedi, Kush R. Varshney, Dennis Wei, Shalisha Witherspooon, Marcel Zalmanovici
Large language models (LLMs) are susceptible to a variety of risks, from non-faithful output to biased and toxic generations.
1 code implementation • 20 Feb 2024 • Maohao Shen, Subhro Das, Kristjan Greenewald, Prasanna Sattigeri, Gregory Wornell, Soumya Ghosh
Addressing these challenges, we propose THERMOMETER, a calibration approach tailored to LLMs.
1 code implementation • 9 Feb 2024 • Maohao Shen, J. Jon Ryu, Soumya Ghosh, Yuheng Bu, Prasanna Sattigeri, Subhro Das, Gregory W. Wornell
This paper questions the effectiveness of a modern predictive uncertainty quantification approach, called \emph{evidential deep learning} (EDL), in which a single neural network model is trained to learn a meta distribution over the predictive distribution by minimizing a specific objective function.
no code implementations • 4 Oct 2023 • Jiri Navratil, Benjamin Elder, Matthew Arnold, Soumya Ghosh, Prasanna Sattigeri
Accurate quantification of model uncertainty has long been recognized as a fundamental requirement for trusted AI.
1 code implementation • 30 Apr 2023 • Maohao Shen, Soumya Ghosh, Prasanna Sattigeri, Subhro Das, Yuheng Bu, Gregory Wornell
Due to privacy or commercial constraints, large pre-trained language models (PLMs) are often offered as black-box APIs.
1 code implementation • 14 Dec 2022 • Maohao Shen, Yuheng Bu, Prasanna Sattigeri, Soumya Ghosh, Subhro Das, Gregory Wornell
It is known that neural networks have the problem of being over-confident when directly using the output label distribution to generate uncertainty measures.
no code implementations • 13 Dec 2022 • Prasanna Sattigeri, Soumya Ghosh, Inkit Padhi, Pierre Dognin, Kush R. Varshney
The dropping of training points is done in principle, but in practice does not require the model to be refit.
no code implementations • 1 Dec 2022 • Sameer K. Deshpande, Soumya Ghosh, Tin D. Nguyen, Tamara Broderick
Test log-likelihood is commonly used to compare different models of the same data or different approximate inference algorithms for fitting the same probabilistic model.
no code implementations • 13 Jun 2021 • Meet P. Vadera, Soumya Ghosh, Kenney Ng, Benjamin M. Marlin
Bayesian decision theory provides an elegant framework for acting optimally under uncertainty when tractable posterior distributions are available.
no code implementations • 11 Jun 2021 • William T. Stephenson, Soumya Ghosh, Tin D. Nguyen, Mikhail Yurochkin, Sameer K. Deshpande, Tamara Broderick
We demonstrate in both synthetic and real-world examples that decisions made with a GP can exhibit non-robustness to kernel choice, even when prior draws are qualitatively interchangeable to a user.
1 code implementation • 2 Jun 2021 • Soumya Ghosh, Q. Vera Liao, Karthikeyan Natesan Ramamurthy, Jiri Navratil, Prasanna Sattigeri, Kush R. Varshney, Yunfeng Zhang
In this paper, we describe an open source Python toolkit named Uncertainty Quantification 360 (UQ360) for the uncertainty quantification of AI models.
1 code implementation • 1 Jun 2021 • Jiri Navratil, Benjamin Elder, Matthew Arnold, Soumya Ghosh, Prasanna Sattigeri
Accurate quantification of model uncertainty has long been recognized as a fundamental requirement for trusted AI.
no code implementations • 18 Dec 2020 • Siddharth Biswal, Soumya Ghosh, Jon Duke, Bradley Malin, Walter Stewart, Jimeng Sun
De-identified EHRs do not adequately address the needs of health systems, as de-identified data are susceptible to re-identification and its volume is also limited.
1 code implementation • ICML 2020 • Sebastian Claici, Mikhail Yurochkin, Soumya Ghosh, Justin Solomon
Our algorithm relies on a mean field assumption for both the fused model and the individual dataset posteriors and proceeds using a simple assign-and-average approach.
1 code implementation • NeurIPS 2020 • Soumya Ghosh, William T. Stephenson, Tin D. Nguyen, Sameer K. Deshpande, Tamara Broderick
But this existing ACV work is restricted to simpler models by the assumptions that (i) data across CV folds are independent and (ii) an exact initial model fit is available.
1 code implementation • NeurIPS 2019 • Mikhail Yurochkin, Mayank Agarwal, Soumya Ghosh, Kristjan Greenewald, Trong Nghia Hoang
We consider the problem of aggregating models learned from sequestered, possibly heterogeneous datasets.
no code implementations • 25 Sep 2019 • Joe Davison, Kristen A. Severson, Soumya Ghosh
A significant body of recent work has examined variational autoencoders as a powerful approach for tasks which involve modeling the distribution of complex data such as images and text.
no code implementations • 24 Jun 2019 • Jiayu Yao, Weiwei Pan, Soumya Ghosh, Finale Doshi-Velez
Bayesian Neural Networks (BNNs) place priors over the parameters in a neural network.
1 code implementation • 28 May 2019 • Mikhail Yurochkin, Mayank Agarwal, Soumya Ghosh, Kristjan Greenewald, Trong Nghia Hoang, Yasaman Khazaeni
In federated learning problems, data is scattered across different servers and exchanging or pooling it is often impractical or prohibited.
no code implementations • ICLR 2019 • Mikhail Yurochkin, Mayank Agarwal, Soumya Ghosh, Kristjan Greenewald, Nghia Hoang, Yasaman Khazaeni
In federated learning problems, data is scattered across different servers and exchanging or pooling it is often impractical or prohibited.
no code implementations • 26 Apr 2019 • Bum Chul Kwon, Vibha Anand, Kristen A Severson, Soumya Ghosh, Zhaonan Sun, Brigitte I Frohnert, Markus Lundgren, Kenney Ng
Clinical researchers use disease progression models to understand patient status and characterize progression patterns from longitudinal health records.
no code implementations • 16 Nov 2018 • Melanie F. Pradier, Weiwei Pan, Jiayu Yao, Soumya Ghosh, Finale Doshi-Velez
As machine learning systems get widely adopted for high-stake decisions, quantifying uncertainty over predictions becomes crucial.
1 code implementation • 14 Nov 2018 • Kristen Severson, Soumya Ghosh, Kenney Ng
Here, we present a probabilistic model for dimensionality reduction to discover signal that is enriched in the target dataset relative to the background dataset.
2 code implementations • ICML 2018 • Soumya Ghosh, Jiayu Yao, Finale Doshi-Velez
Bayesian Neural Networks (BNNs) have recently received increasing attention for their ability to provide well-calibrated posterior uncertainties.
no code implementations • 19 Feb 2018 • Bin Liu, Ying Li, Soumya Ghosh, Zhaonan Sun, Kenney Ng, Jianying Hu
The proposed method is favorable for healthcare applications because in additional to improved prediction performance, relationships among the different risks and risk factors are also identified.
no code implementations • CVPR 2017 • Ajjen Joshi, Soumya Ghosh, Margrit Betke, Stan Sclaroff, Hanspeter Pfister
Leveraging recent work on learning Bayesian neural networks, we build fast, scalable algorithms for inferring the posterior distribution over all network weights in the hierarchy.
1 code implementation • 29 May 2017 • Soumya Ghosh, Finale Doshi-Velez
Bayesian Neural Networks (BNNs) have recently received increasing attention for their ability to provide well-calibrated posterior uncertainties.
2 code implementations • NeurIPS 2012 • Soumya Ghosh, Matthew Loper, Erik B. Sudderth, Michael J. Black
We develop a method for discovering the parts of an articulated object from aligned meshes capturing various three-dimensional (3D) poses.
no code implementations • NeurIPS 2011 • Soumya Ghosh, Andrei B. Ungureanu, Erik B. Sudderth, David M. Blei
The distance dependent Chinese restaurant process (ddCRP) was recently introduced to accommodate random partitions of non-exchangeable data.