2 code implementations • 20 Aug 2024 • Vitus Benson, Ana Bastos, Christian Reimers, Alexander J. Winkler, Fanny Yang, Markus Reichstein
Accurately describing the distribution of CO$_2$ in the atmosphere with atmospheric tracer transport models is essential for greenhouse gas monitoring and verification support systems to aid implementation of international climate agreements.
no code implementations • 29 Jul 2024 • Javier Abad, Konstantin Donhauser, Francesco Pinto, Fanny Yang
The risk of language models unintentionally reproducing copyrighted material from their training data has led to the development of various protective measures.
no code implementations • 22 Jul 2024 • Daniil Dmitriev, Rares-Darius Buhai, Stefan Tiegel, Alexander Wolters, Gleb Novikov, Amartya Sanyal, David Steurer, Fanny Yang
We study the problem of estimating the means of well-separated mixtures when an adversary may add arbitrary outliers.
1 code implementation • 29 Apr 2024 • Piersilvio De Bartolomeis, Javier Abad, Konstantin Donhauser, Fanny Yang
Randomized trials are considered the gold standard for making informed decisions in medicine, yet they often lack generalizability to the patient populations in clinical practice.
1 code implementation • 31 Jan 2024 • Konstantin Donhauser, Javier Abad, Neha Hulkund, Fanny Yang
We present a novel approach for differentially private data synthesis of protected tabular datasets, a relevant task in highly sensitive domains such as healthcare and government.
2 code implementations • 6 Dec 2023 • Piersilvio De Bartolomeis, Javier Abad, Konstantin Donhauser, Fanny Yang
Further, we show how our lower bound can correctly identify the absence and presence of unobserved confounding in a real-world setting.
no code implementations • NeurIPS 2023 • Alexandru Ţifrea, Gizem Yüce, Amartya Sanyal, Fanny Yang
Prior works have shown that semi-supervised learning algorithms can leverage unlabeled data to improve over the labeled sample complexity of supervised learning (SL) algorithms.
no code implementations • 12 Jun 2023 • Piersilvio De Bartolomeis, Jacob Clarysse, Amartya Sanyal, Fanny Yang
In this paper, we systematically compare the standard and robust error of these two robust training paradigms across multiple computer vision tasks.
1 code implementation • 6 Jun 2023 • Francesco Pinto, Yaxi Hu, Fanny Yang, Amartya Sanyal
In Semi-Supervised Semi-Private (SP) learning, the learner has access to both public unlabelled and private labelled data.
1 code implementation • 18 Jan 2023 • Michael Aerni, Marco Milanta, Konstantin Donhauser, Fanny Yang
Classical wisdom suggests that estimators should avoid fitting noise to achieve good generalization.
no code implementations • 7 Dec 2022 • Stefan Stojanovic, Konstantin Donhauser, Fanny Yang
In particular, for the noiseless setting, we prove tight upper and lower bounds for the prediction error that match existing rates of order $\frac{\|w^*\|_1^{2/3}}{n^{1/3}}$ for general ground truths.
no code implementations • 1 Dec 2022 • Alexandru Tifrea, Jacob Clarysse, Fanny Yang
It is widely believed that given the same labeling budget, active learning (AL) algorithms like margin-based active learning achieve better predictive performance than passive learning (PL), albeit at a higher computational cost.
no code implementations • 8 Jun 2022 • Amartya Sanyal, Yaxi Hu, Fanny Yang
As machine learning algorithms are deployed on sensitive data in critical decision making processes, it is becoming increasingly important that they are also private and fair.
2 code implementations • 1 Apr 2022 • Armeen Taeb, Nicolo Ruggeri, Carina Schnuck, Fanny Yang
In safety-critical applications, practitioners are reluctant to trust neural networks when no interpretable explanations are available.
1 code implementation • 7 Mar 2022 • Konstantin Donhauser, Nicolo Ruggeri, Stefan Stojanovic, Fanny Yang
Good generalization performance on high-dimensional data crucially hinges on a simple structure of the ground truth and a corresponding strong inductive bias of the estimator.
no code implementations • 3 Mar 2022 • Jacob Clarysse, Julia Hörmann, Fanny Yang
Machine learning classifiers with high test accuracy often perform poorly under adversarial attacks.
1 code implementation • 10 Nov 2021 • Guillaume Wang, Konstantin Donhauser, Fanny Yang
We provide matching upper and lower bounds of order $\sigma^2/\log(d/n)$ for the prediction error of the minimum $\ell_1$-norm interpolator, a. k. a.
no code implementations • 9 Sep 2021 • Andrii Zadaianchuk, Georg Martius, Fanny Yang
We propose a novel self-supervised agent that estimates relations between environment components and uses them to independently control different parts of the environment state.
2 code implementations • NeurIPS 2021 • Konstantin Donhauser, Alexandru Ţifrea, Michael Aerni, Reinhard Heckel, Fanny Yang
Numerous recent works show that overparameterization implicitly reduces variance for min-norm interpolators and max-margin classifiers.
1 code implementation • ICML Workshop AML 2021 • Konstantin Donhauser, Alexandru Tifrea, Michael Aerni, Reinhard Heckel, Fanny Yang
Numerous recent works show that overparameterization implicitly reduces variance, suggesting vanishing benefits for explicit regularization in high dimensions.
1 code implementation • 9 Apr 2021 • Konstantin Donhauser, Mingqi Wu, Fanny Yang
Kernel ridge regression is well-known to achieve minimax optimal rates in low-dimensional settings.
1 code implementation • 10 Dec 2020 • Alexandru Ţifrea, Eric Stavarache, Fanny Yang
Deep neural networks often predict samples with high confidence even when they come from unseen classes and should instead be flagged for expert evaluation.
Ranked #1 on Out-of-Distribution Detection on CIFAR-10 vs CIFAR-10.1 (using extra training data)
no code implementations • 28 Sep 2020 • Alexandru Țifrea, Eric Petru Stavarache, Fanny Yang
This paper studies how such ``hard'' OOD scenarios can benefit from tuning the detection method after observing a batch of the test data.
Out-of-Distribution Detection Out of Distribution (OOD) Detection
1 code implementation • ICML 2020 • Aditi Raghunathan, Sang Michael Xie, Fanny Yang, John Duchi, Percy Liang
In this work, we precisely characterize the effect of augmentation on the standard error in linear regression when the optimal linear predictor has zero standard and robust error.
no code implementations • 25 Sep 2019 • Sang Michael Xie*, Aditi Raghunathan*, Fanny Yang, John C. Duchi, Percy Liang
Empirically, data augmentation sometimes improves and sometimes hurts test error, even when only adding points with labels from the true conditional distribution that the hypothesis class is expressive enough to fit.
no code implementations • NeurIPS 2019 • Fanny Yang, Zuowen Wang, Christina Heinze-Deml
This work provides theoretical and empirical evidence that invariance-inducing regularizers can increase predictive accuracy for worst-case spatial transformations (spatial robustness).
no code implementations • ICML Workshop Deep_Phenomen 2019 • Aditi Raghunathan, Sang Michael Xie, Fanny Yang, John C. Duchi, Percy Liang
While adversarial training can improve robust accuracy (against an adversary), it sometimes hurts standard accuracy (when there is no adversary).
2 code implementations • ICLR 2019 • Kamyar Azizzadenesheli, Anqi Liu, Fanny Yang, Animashree Anandkumar
We derive a generalization bound for the classifier on the target domain which is independent of the (ambient) data dimensions, and instead only depends on the complexity of the function class.
1 code implementation • NeurIPS 2017 • Aaditya Ramdas, Fanny Yang, Martin J. Wainwright, Michael. I. Jordan
In the online multiple testing problem, p-values corresponding to different null hypotheses are observed one by one, and the decision of whether or not to reject the current hypothesis must be made immediately, after which the next p-value is observed.
no code implementations • NeurIPS 2017 • Yuting Wei, Fanny Yang, Martin J. Wainwright
Early stopping of iterative algorithms is a widely-used form of regularization in statistics, commonly used in conjunction with boosting and related gradient-type algorithms.
1 code implementation • NeurIPS 2017 • Fanny Yang, Aaditya Ramdas, Kevin Jamieson, Martin J. Wainwright
We propose an alternative framework to existing setups for controlling false alarms when multiple A/B tests are run over time.
no code implementations • 27 Dec 2015 • Fanny Yang, Sivaraman Balakrishnan, Martin J. Wainwright
By exploiting this characterization, we provide non-asymptotic finite sample guarantees on the Baum-Welch updates, guaranteeing geometric convergence to a small ball of radius on the order of the minimax rate around a global optimum.