no code implementations • ICML 2020 • Debarghya Mukherjee, Mikhail Yurochkin, Moulinath Banerjee, Yuekai Sun
Individual fairness was proposed to address some of the shortcomings of group fairness.
no code implementations • 20 Apr 2024 • Seamus Somerstep, Yuekai Sun, Ya'acov Ritov
Motivated by equilibrium models of labor markets, we develop a formulation of causal strategic classification in which strategic agents can directly manipulate their outcomes.
no code implementations • 7 Mar 2024 • Lilian Ngweta, Mayank Agarwal, Subha Maity, Alex Gittens, Yuekai Sun, Mikhail Yurochkin
Large Language Models (LLMs) need to be aligned with human expectations to ensure their safety and utility in most applications.
2 code implementations • 22 Feb 2024 • Felipe Maia Polo, Lucas Weber, Leshem Choshen, Yuekai Sun, Gongjun Xu, Mikhail Yurochkin
The versatility of large language models (LLMs) led to the creation of diverse benchmarks that thoroughly test a variety of language models' abilities.
no code implementations • 7 Dec 2023 • Felipe Maia Polo, Mikhail Yurochkin, Moulinath Banerjee, Subha Maity, Yuekai Sun
We develop methods for estimating Fr\'echet bounds on (possibly high-dimensional) distribution classes in which some variables are continuous-valued.
no code implementations • 2 Oct 2023 • Hongyi Wang, Felipe Maia Polo, Yuekai Sun, Souvik Kundu, Eric Xing, Mikhail Yurochkin
Training AI models that generalize across tasks and domains has long been among the open problems driving AI research.
1 code implementation • 2 Oct 2023 • Subha Maity, Mayank Agarwal, Mikhail Yurochkin, Yuekai Sun
In this paper, we demonstrate that contrastive learning (CL), a popular variant of SSL, tends to collapse representations of minority groups with certain majority groups.
no code implementations • 27 Sep 2023 • Tal Shnitzer, Anthony Ou, Mírian Silva, Kate Soule, Yuekai Sun, Justin Solomon, Neil Thompson, Mikhail Yurochkin
There is a rapidly growing number of open-source Large Language Models (LLMs) and benchmark datasets to compare them.
1 code implementation • NeurIPS 2023 • Felipe Maia Polo, Yuekai Sun, Moulinath Banerjee
Conditional independence (CI) testing is a fundamental and challenging task in modern statistics and machine learning.
1 code implementation • 1 May 2023 • Felix Petersen, Tobias Sutter, Christian Borgelt, Dongsung Huh, Hilde Kuehne, Yuekai Sun, Oliver Deussen
We present ISAAC (Input-baSed ApproximAte Curvature), a novel method that conditions the gradient using selected second-order information and has an asymptotically vanishing computational overhead, assuming a batch size smaller than the number of neurons.
1 code implementation • 20 Feb 2023 • Lilian Ngweta, Subha Maity, Alex Gittens, Yuekai Sun, Mikhail Yurochkin
Learning visual representations with interpretable features, i. e., disentangled representations, remains a challenging problem.
no code implementations • 15 Jan 2023 • Songkai Xue, Yuekai Sun, Mikhail Yurochkin
We consider the task of training machine learning models with data-dependent constraints.
1 code implementation • 7 Jun 2022 • Subha Maity, Saptarshi Roy, Songkai Xue, Mikhail Yurochkin, Yuekai Sun
The benefits of overparameterization for the overall performance of modern machine learning (ML) models are well known.
1 code implementation • 26 May 2022 • Subha Maity, Debarghya Mukherjee, Moulinath Banerjee, Yuekai Sun
Time-varying stochastic optimization problems frequently arise in machine learning practice (e. g. gradual domain shift, object tracking, strategic classification).
1 code implementation • 26 May 2022 • Subha Maity, Mikhail Yurochkin, Moulinath Banerjee, Yuekai Sun
However, it is conceivable that the training data can be reweighted to be more representative of the new (target) task.
no code implementations • 1 May 2022 • Debarghya Mukherjee, Felix Petersen, Mikhail Yurochkin, Yuekai Sun
In this paper, we leverage this connection between algorithmic fairness and distribution shifts to show that algorithmic fairness interventions can help ML models overcome distribution shifts, and that domain adaptation methods (for overcoming distribution shifts) can mitigate algorithmic biases.
no code implementations • 13 Apr 2022 • Laura Niss, Yuekai Sun, Ambuj Tewari
Sampling biases in training data are a major source of algorithmic biases in machine learning systems.
1 code implementation • NeurIPS 2021 • Felix Petersen, Debarghya Mukherjee, Yuekai Sun, Mikhail Yurochkin
In this work, we propose general post-processing algorithms for individual fairness (IF).
no code implementations • NeurIPS 2021 • Mayank Agarwal, Mikhail Yurochkin, Yuekai Sun
Meta-learning algorithms are widely used for few-shot learning.
no code implementations • ICLR 2021 • Alexander Vargo, Fan Zhang, Mikhail Yurochkin, Yuekai Sun
Gradient boosting is a popular method for machine learning from tabular data, which arise often in applications where algorithmic fairness is a concern.
1 code implementation • ICLR 2021 • Subha Maity, Songkai Xue, Mikhail Yurochkin, Yuekai Sun
As we rely on machine learning (ML) models to make more consequential decisions, the issue of ML models perpetuating or even exacerbating undesirable historical biases (e. g., gender and racial biases) has come to the fore of the public's attention.
no code implementations • 19 Mar 2021 • Amanda Bower, Hamid Eftekhari, Mikhail Yurochkin, Yuekai Sun
We develop an algorithm to train individually fair learning-to-rank (LTR) models.
no code implementations • 1 Jan 2021 • Debarghya Mukherjee, Aritra Guha, Justin Solomon, Yuekai Sun, Mikhail Yurochkin
In light of recent advances in solving the OT problem, OT distances are widely used as loss functions in minimum distance estimation.
no code implementations • ICLR 2021 • Amanda Bower, Hamid Eftekhari, Mikhail Yurochkin, Yuekai Sun
We develop an algorithm to train individually fair learning-to-rank (LTR) models.
no code implementations • NeurIPS 2021 • Subha Maity, Debarghya Mukherjee, Mikhail Yurochkin, Yuekai Sun
Many instances of algorithmic bias are caused by subpopulation shifts.
no code implementations • 28 Sep 2020 • Subha Maity, Debarghya Mukherjee, Mikhail Yurochkin, Yuekai Sun
If the algorithmic biases in an ML model are due to sampling biases in the training data, then enforcing algorithmic fairness may improve the performance of the ML model on unbiased test data.
no code implementations • ICLR 2021 • Mikhail Yurochkin, Yuekai Sun
In this paper, we cast fair machine learning as invariant machine learning.
no code implementations • 19 Jun 2020 • Debarghya Mukherjee, Mikhail Yurochkin, Moulinath Banerjee, Yuekai Sun
Individual fairness is an intuitive definition of algorithmic fairness that addresses some of the drawbacks of group fairness.
no code implementations • 23 Mar 2020 • Subha Maity, Yuekai Sun, Moulinath Banerjee
We study the minimax rates of the label shift problem in non-parametric classification.
no code implementations • 11 Mar 2020 • Songkai Xue, Mikhail Yurochkin, Yuekai Sun
We consider the task of auditing ML models for individual bias/unfairness.
1 code implementation • ICLR 2020 • Hongyi Wang, Mikhail Yurochkin, Yuekai Sun, Dimitris Papailiopoulos, Yasaman Khazaeni
Federated learning allows edge devices to collaboratively learn a shared model while keeping the training data on device, decoupling the ability to do model training from the need to store the data in the cloud.
1 code implementation • 26 Dec 2019 • Subha Maity, Yuekai Sun, Moulinath Banerjee
We consider the task of meta-analysis in high-dimensional settings in which the data sources are similar but non-identical.
2 code implementations • ICLR 2020 • Mikhail Yurochkin, Amanda Bower, Yuekai Sun
We consider training machine learning models that are fair in the sense that their performance is invariant under certain sensitive perturbations to the inputs.
1 code implementation • 27 May 2019 • Mikhail Yurochkin, Aritra Guha, Yuekai Sun, XuanLong Nguyen
We propose Dirichlet Simplex Nest, a class of probabilistic models suitable for a variety of data types, and develop fast and provably accurate inference algorithms by accounting for the model's convex geometry and low dimensional simplicial structure.
no code implementations • 7 Apr 2019 • Roger Fan, Byoungwook Jang, Yuekai Sun, Shuheng Zhou
Estimating conditional dependence graphs and precision matrices are some of the most common problems in modern statistics and machine learning.
2 code implementations • 11 Oct 2017 • Laurent Heirendt, Sylvain Arreckx, Thomas Pfau, Sebastián N. Mendoza, Anne Richelle, Almut Heinken, Hulda S. Haraldsdóttir, Jacek Wachowiak, Sarah M. Keating, Vanja Vlasov, Stefania Magnusdóttir, Chiam Yu Ng, German Preciat, Alise Žagare, Siu H. J. Chan, Maike K. Aurich, Catherine M. Clancy, Jennifer Modamio, John T. Sauls, Alberto Noronha, Aarash Bordbar, Benjamin Cousins, Diana C. El Assal, Luis V. Valcarcel, Iñigo Apaolaza, Susan Ghaderi, Masoud Ahookhosh, Marouen Ben Guebila, Andrejs Kostromins, Nicolas Sompairac, Hoai M. Le, Ding Ma, Yuekai Sun, Lin Wang, James T. Yurkovich, Miguel A. P. Oliveira, Phan T. Vuong, Lemmer P. El Assal, Inna Kuperstein, Andrei Zinovyev, H. Scott Hinton, William A. Bryant, Francisco J. Aragón Artacho, Francisco J. Planes, Egils Stalidzans, Alejandro Maass, Santosh Vempala, Michael Hucka, Michael A. Saunders, Costas D. Maranas, Nathan E. Lewis, Thomas Sauter, Bernhard Ø. Palsson, Ines Thiele, Ronan M. T. Fleming
This protocol can be adapted for the generation and analysis of a constraint-based model in a wide variety of molecular systems biology scenarios.
no code implementations • 28 Aug 2017 • Xuanqing Liu, Cho-Jui Hsieh, Jason D. Lee, Yuekai Sun
We propose a fast proximal Newton-type algorithm for minimizing regularized finite sums that returns an $\epsilon$-suboptimal point in $\tilde{\mathcal{O}}(d(n + \sqrt{\kappa d})\log(\frac{1}{\epsilon}))$ FLOPS, where $n$ is number of samples, $d$ is feature dimension, and $\kappa$ is the condition number.
no code implementations • 26 Jun 2017 • Ya'acov Ritov, Yuekai Sun, Ruofei Zhao
We identify conditional parity as a general notion of non-discrimination in machine learning.
no code implementations • NeurIPS 2016 • Jiyan Yang, Michael W. Mahoney, Michael Saunders, Yuekai Sun
Most existing approaches to distributed sparse regression assume the data is partitioned by samples.
no code implementations • NeurIPS 2015 • Jason D. Lee, Yuekai Sun, Jonathan E. Taylor
Biclustering (also known as submatrix localization) is a problem of high practical relevance in exploratory analysis of high-dimensional data.
no code implementations • 14 Mar 2015 • Jason D. Lee, Yuekai Sun, Qiang Liu, Jonathan E. Taylor
We devise a one-shot approach to distributed sparse regression in the high-dimensional setting.
no code implementations • NeurIPS 2013 • Jason D. Lee, Yuekai Sun, Jonathan E. Taylor
Penalized M-estimators are used in diverse areas of science and engineering to fit high-dimensional models with some low-dimensional structure.
no code implementations • 25 Nov 2013 • Jason D. Lee, Dennis L. Sun, Yuekai Sun, Jonathan E. Taylor
We develop a general approach to valid inference after model selection.
no code implementations • 11 Nov 2013 • Yuekai Sun, Stratis Ioannidis, Andrea Montanari
We consider a discriminative learning (regression) problem, whereby the regression function is a convex combination of k linear classifiers.
no code implementations • 31 May 2013 • Jason D. Lee, Yuekai Sun, Jonathan E. Taylor
Regularized M-estimators are used in diverse areas of science and engineering to fit high-dimensional models with some low-dimensional structure.
1 code implementation • 7 Jun 2012 • Jason D. Lee, Yuekai Sun, Michael A. Saunders
We generalize Newton-type methods for minimizing smooth functions to handle a sum of two convex functions: a smooth function and a nonsmooth function with a simple proximal mapping.