no code implementations • ICML 2020 • Debarghya Mukherjee, Mikhail Yurochkin, Moulinath Banerjee, Yuekai Sun
Individual fairness was proposed to address some of the shortcomings of group fairness.
no code implementations • 1 Mar 2023 • Zahra Ashktorab, Benjamin Hoover, Mayank Agarwal, Casey Dugan, Werner Geyer, Hao Bang Yang, Mikhail Yurochkin
While several toolkits exist to aid machine learning practitioners in addressing fairness issues, little is known about the strategies practitioners employ to evaluate model fairness and what factors influence their assessment, particularly in the context of text classification.
1 code implementation • 20 Feb 2023 • Lilian Ngweta, Subha Maity, Alex Gittens, Yuekai Sun, Mikhail Yurochkin
Learning visual representations with interpretable features, i. e., disentangled representations, remains a challenging problem.
no code implementations • 15 Jan 2023 • Songkai Xue, Yuekai Sun, Mikhail Yurochkin
We consider the task of training machine learning models with data-dependent constraints.
1 code implementation • 24 Oct 2022 • Lingxiao Li, Qiang Liu, Anna Korba, Mikhail Yurochkin, Justin Solomon
These energies rely on mollifier functions -- smooth approximations of the Dirac delta originated from PDE theory.
1 code implementation • 13 Oct 2022 • Yuchen Zeng, Kristjan Greenewald, Kangwook Lee, Justin Solomon, Mikhail Yurochkin
Traditional machine learning models focus on achieving good performance on the overall training distribution, but they often underperform on minority groups.
1 code implementation • 7 Jun 2022 • Subha Maity, Saptarshi Roy, Songkai Xue, Mikhail Yurochkin, Yuekai Sun
The benefits of overparameterization for the overall performance of modern machine learning (ML) models are well known.
1 code implementation • 26 May 2022 • Subha Maity, Mikhail Yurochkin, Moulinath Banerjee, Yuekai Sun
However, it is conceivable that the training data can be reweighted to be more representative of the new (target) task.
no code implementations • 1 May 2022 • Debarghya Mukherjee, Felix Petersen, Mikhail Yurochkin, Yuekai Sun
In this paper, we leverage this connection between algorithmic fairness and distribution shifts to show that algorithmic fairness interventions can help ML models overcome distribution shifts, and that domain adaptation methods (for overcoming distribution shifts) can mitigate algorithmic biases.
1 code implementation • 3 Feb 2022 • Tal Shnitzer, Mikhail Yurochkin, Kristjan Greenewald, Justin Solomon
We use manifold learning to compare the intrinsic geometric structures of different datasets by comparing their diffusion operators, symmetric positive-definite (SPD) matrices that relate to approximations of the continuous Laplace-Beltrami operator from discrete samples.
no code implementations • 29 Jan 2022 • Rickard Brüel-Gabrielsson, Mikhail Yurochkin, Justin Solomon
As a conservative alternative, we use positional encodings to expand receptive fields to $r$-hop neighborhoods.
1 code implementation • 28 Jan 2022 • Lingxiao Li, Noam Aigerman, Vladimir G. Kim, Jiajin Li, Kristjan Greenewald, Mikhail Yurochkin, Justin Solomon
We present an end-to-end method to learn the proximal operator of a family of training problems so that multiple local minima can be quickly obtained from initial guesses by iterating the learned operator, emulating the proximal-point algorithm that has fast convergence.
no code implementations • NeurIPS 2021 • Mayank Agarwal, Mikhail Yurochkin, Yuekai Sun
Meta-learning algorithms are widely used for few-shot learning.
1 code implementation • NeurIPS 2021 • Felix Petersen, Debarghya Mukherjee, Yuekai Sun, Mikhail Yurochkin
In this work, we propose general post-processing algorithms for individual fairness (IF).
no code implementations • 29 Sep 2021 • Kristjan Greenewald, Anming Gu, Mikhail Yurochkin, Justin Solomon, Edward Chien
To better leverage the structure of the data, we extend mixup to $k$-mixup by perturbing $k$-batches of training points in the direction of other $k$-batches using displacement interpolation, i. e. interpolation under the Wasserstein metric.
no code implementations • 29 Sep 2021 • Felix Petersen, Christian Borgelt, Mikhail Yurochkin, Hilde Kuehne, Oliver Deussen
We propose a new approach to propagating probability distributions through neural networks.
no code implementations • Findings (ACL) 2022 • Ioana Baldini, Dennis Wei, Karthikeyan Natesan Ramamurthy, Mikhail Yurochkin, Moninder Singh
Through the analysis of more than a dozen pretrained language models of varying sizes on two toxic text classification tasks (English), we demonstrate that focusing on accuracy measures alone can lead to models with wide variation in fairness characteristics.
no code implementations • 11 Jun 2021 • William T. Stephenson, Soumya Ghosh, Tin D. Nguyen, Mikhail Yurochkin, Sameer K. Deshpande, Tamara Broderick
We demonstrate in both synthetic and real-world examples that decisions made with a GP can exhibit non-robustness to kernel choice, even when prior draws are qualitatively interchangeable to a user.
no code implementations • 5 Jun 2021 • Kristjan Greenewald, Anming Gu, Mikhail Yurochkin, Justin Solomon, Edward Chien
Mixup is a popular regularization technique for training deep neural networks that can improve generalization and increase adversarial robustness.
no code implementations • ICLR 2021 • Alexander Vargo, Fan Zhang, Mikhail Yurochkin, Yuekai Sun
Gradient boosting is a popular method for machine learning from tabular data, which arise often in applications where algorithmic fairness is a concern.
1 code implementation • ICLR 2021 • Subha Maity, Songkai Xue, Mikhail Yurochkin, Yuekai Sun
As we rely on machine learning (ML) models to make more consequential decisions, the issue of ML models perpetuating or even exacerbating undesirable historical biases (e. g., gender and racial biases) has come to the fore of the public's attention.
no code implementations • 19 Mar 2021 • Amanda Bower, Hamid Eftekhari, Mikhail Yurochkin, Yuekai Sun
We develop an algorithm to train individually fair learning-to-rank (LTR) models.
no code implementations • ICLR 2021 • Amanda Bower, Hamid Eftekhari, Mikhail Yurochkin, Yuekai Sun
We develop an algorithm to train individually fair learning-to-rank (LTR) models.
no code implementations • 1 Jan 2021 • Debarghya Mukherjee, Aritra Guha, Justin Solomon, Yuekai Sun, Mikhail Yurochkin
In light of recent advances in solving the OT problem, OT distances are widely used as loss functions in minimum distance estimation.
no code implementations • NeurIPS 2021 • Subha Maity, Debarghya Mukherjee, Mikhail Yurochkin, Yuekai Sun
Many instances of algorithmic bias are caused by subpopulation shifts.
no code implementations • ICLR Workshop LLD 2019 • Sohini Upadhyay, Mikhail Yurochkin, Mayank Agarwal, Yasaman Khazaeni, DjallelBouneffouf
We formulate a new problem at the intersectionof semi-supervised learning and contextual bandits, motivated by several applications including clini-cal trials and ad recommendations.
no code implementations • 28 Sep 2020 • Subha Maity, Debarghya Mukherjee, Mikhail Yurochkin, Yuekai Sun
If the algorithmic biases in an ML model are due to sampling biases in the training data, then enforcing algorithmic fairness may improve the performance of the ML model on unbiased test data.
1 code implementation • NeurIPS 2020 • Lingxiao Li, Aude Genevay, Mikhail Yurochkin, Justin Solomon
Leveraging a new dual formulation for the regularized Wasserstein barycenter problem, we introduce a stochastic algorithm that constructs a continuous approximation of the barycenter.
1 code implementation • 22 Jul 2020 • Heiko Ludwig, Nathalie Baracaldo, Gegi Thomas, Yi Zhou, Ali Anwar, Shashank Rajamoni, Yuya Ong, Jayaram Radhakrishnan, Ashish Verma, Mathieu Sinn, Mark Purcell, Ambrish Rawat, Tran Minh, Naoise Holohan, Supriyo Chakraborty, Shalisha Whitherspoon, Dean Steuer, Laura Wynter, Hifaz Hassan, Sean Laguna, Mikhail Yurochkin, Mayank Agarwal, Ebube Chuba, Annie Abay
Federated Learning (FL) is an approach to conduct machine learning without centralizing training data in a single place, for reasons of privacy, confidentiality or data volume.
1 code implementation • ICML 2020 • Sebastian Claici, Mikhail Yurochkin, Soumya Ghosh, Justin Solomon
Our algorithm relies on a mean field assumption for both the fused model and the individual dataset posteriors and proceeds using a simple assign-and-average approach.
no code implementations • ICLR 2021 • Mikhail Yurochkin, Yuekai Sun
In this paper, we cast fair machine learning as invariant machine learning.
no code implementations • 19 Jun 2020 • Debarghya Mukherjee, Mikhail Yurochkin, Moulinath Banerjee, Yuekai Sun
Individual fairness is an intuitive definition of algorithmic fairness that addresses some of the drawbacks of group fairness.
no code implementations • 11 Mar 2020 • Songkai Xue, Mikhail Yurochkin, Yuekai Sun
We consider the task of auditing ML models for individual bias/unfairness.
1 code implementation • ICLR 2020 • Hongyi Wang, Mikhail Yurochkin, Yuekai Sun, Dimitris Papailiopoulos, Yasaman Khazaeni
Federated learning allows edge devices to collaboratively learn a shared model while keeping the training data on device, decoupling the ability to do model training from the need to store the data in the cloud.
1 code implementation • NeurIPS 2019 • Pierre Monteiller, Sebastian Claici, Edward Chien, Farzaneh Mirzazadeh, Justin Solomon, Mikhail Yurochkin
Label switching is a phenomenon arising in mixture model posterior inference that prevents one from meaningfully assessing posterior statistics using standard Monte Carlo procedures.
1 code implementation • NeurIPS 2019 • Mikhail Yurochkin, Mayank Agarwal, Soumya Ghosh, Kristjan Greenewald, Trong Nghia Hoang
We consider the problem of aggregating models learned from sequestered, possibly heterogeneous datasets.
1 code implementation • 19 Sep 2019 • Viet Huynh, Nhat Ho, Nhan Dam, XuanLong Nguyen, Mikhail Yurochkin, Hung Bui, and Dinh Phung
We propose a novel approach to the problem of multilevel clustering, which aims to simultaneously partition data in each group and discover grouping patterns among groups in a potentially large hierarchically structured corpus of data.
2 code implementations • ICLR 2020 • Mikhail Yurochkin, Amanda Bower, Yuekai Sun
We consider training machine learning models that are fair in the sense that their performance is invariant under certain sensitive perturbations to the inputs.
1 code implementation • NeurIPS 2019 • Mikhail Yurochkin, Sebastian Claici, Edward Chien, Farzaneh Mirzazadeh, Justin Solomon
The ability to measure similarity between documents enables intelligent summarization and analysis of large corpora.
1 code implementation • 28 May 2019 • Mikhail Yurochkin, Mayank Agarwal, Soumya Ghosh, Kristjan Greenewald, Trong Nghia Hoang, Yasaman Khazaeni
In federated learning problems, data is scattered across different servers and exchanging or pooling it is often impractical or prohibited.
1 code implementation • 27 May 2019 • Mikhail Yurochkin, Aritra Guha, Yuekai Sun, XuanLong Nguyen
We propose Dirichlet Simplex Nest, a class of probabilistic models suitable for a variety of data types, and develop fast and provably accurate inference algorithms by accounting for the model's convex geometry and low dimensional simplicial structure.
no code implementations • ICLR 2019 • Mikhail Yurochkin, Mayank Agarwal, Soumya Ghosh, Kristjan Greenewald, Nghia Hoang, Yasaman Khazaeni
In federated learning problems, data is scattered across different servers and exchanging or pooling it is often impractical or prohibited.
1 code implementation • NeurIPS 2019 • Mikhail Yurochkin, Zhiwei Fan, Aritra Guha, Paraschos Koutris, XuanLong Nguyen
We develop new models and algorithms for learning the temporal dynamics of the topic polytopes and related geometric objects that arise in topic model based inference.
no code implementations • ICLR 2018 • Mikhail Yurochkin, Dung Thai, Hung Hai Bui, XuanLong Nguyen
In this work we propose a novel approach for learning graph representation of the data using gradients obtained via backpropagation.
1 code implementation • NeurIPS 2017 • Mikhail Yurochkin, Aritra Guha, XuanLong Nguyen
We propose new algorithms for topic modeling when the number of topics is unknown.
1 code implementation • NeurIPS 2017 • Mikhail Yurochkin, XuanLong Nguyen, Nikolaos Vasiloglou
We propose a Bayesian regression method that accounts for multi-way interactions of arbitrary orders among the predictor variables.
1 code implementation • ICML 2017 • Nhat Ho, XuanLong Nguyen, Mikhail Yurochkin, Hung Hai Bui, Viet Huynh, Dinh Phung
We propose a novel approach to the problem of multilevel clustering, which aims to simultaneously partition data in each group and discover grouping patterns among groups in a potentially large hierarchically structured corpus of data.
no code implementations • NeurIPS 2016 • Mikhail Yurochkin, XuanLong Nguyen
We propose a geometric algorithm for topic learning and inference that is built on the convex geometry of topics arising from the Latent Dirichlet Allocation (LDA) model and its nonparametric extensions.