no code implementations • 10 Nov 2024 • Tian Xie, Jifan Zhang, Haoyue Bai, Robert Nowak
Machine learning models deployed in open-world scenarios often encounter unfamiliar conditions and perform poorly in unanticipated situations.
no code implementations • 10 Oct 2024 • Haoyue Bai, Jifan Zhang, Robert Nowak
This paper introduces a novel, integrated approach AHA (Adaptive Human-Assisted OOD learning) to simultaneously address both OOD generalization and detection through a human-assisted framework by labeling data in the wild.
no code implementations • 3 Oct 2024 • Jifan Zhang, Robert Nowak
SIEVE can perform up to 500 filtering operations for the cost of one GPT-4o filtering call.
1 code implementation • 15 Jun 2024 • Jifan Zhang, Lalit Jain, Yang Guo, Jiayi Chen, Kuan Lok Zhou, Siddharth Suresh, Andrew Wagenmaker, Scott Sievert, Timothy Rogers, Kevin Jamieson, Robert Mankoff, Robert Nowak
We present a novel multimodal preference dataset for creative tasks, consisting of over 250 million human ratings on more than 2. 2 million captions, collected through crowdsourcing rating data for The New Yorker's weekly cartoon caption contest over the past eight years.
no code implementations • 7 Jun 2024 • Subhojyoti Mukherjee, Josiah P. Hanna, Qiaomin Xie, Robert Nowak
Interestingly, we show that our algorithm, without the knowledge of the underlying problem structure, can learn a near-optimal policy in-context by leveraging the shared structure across diverse tasks.
no code implementations • 4 Jun 2024 • Subhojyoti Mukherjee, Josiah P. Hanna, Robert Nowak
We then introduce an algorithm SaVeR for this problem that approximates the safe oracle algorithm and bound the finite-sample mean squared error of the algorithm while ensuring it satisfies the safety constraint.
no code implementations • 11 Feb 2024 • Jeongyeol Kwon, Liu Yang, Robert Nowak, Josiah Hanna
Then, our main contributions are two-fold: (a) we demonstrate that the performance of reinforcement learning is strongly correlated with the prediction accuracy of future observations in partially observable environments, and (b) our approach can significantly improve the overall end-to-end approach by preventing high-variance noisy signals from reinforcement learning objectives to influence the representation learning.
no code implementations • 23 Jan 2024 • Nasim Soltani, Jifan Zhang, Batool Salehi, Debashri Roy, Robert Nowak, Kaushik Chowdhury
We evaluate the performance of different active learning algorithms on a publicly available multi-modal dataset with different modalities including image and LiDAR.
no code implementations • 14 Dec 2023 • Shyam Nuggehalli, Jifan Zhang, Lalit Jain, Robert Nowak
Our results demonstrate that DIRECT can save more than 60% of the annotation budget compared to state-of-art active learning algorithms and more than 80% of annotation budget compared to random sampling.
1 code implementation • 21 Nov 2023 • Liu Yang, Kangwook Lee, Robert Nowak, Dimitris Papailiopoulos
Transformers have demonstrated effectiveness in in-context solving data-fitting problems from various (latent) models, as reported by Garg et al.
no code implementations • 4 Sep 2023 • Jeongyeol Kwon, Dohyun Kwon, Stephen Wright, Robert Nowak
When the perturbed lower-level problem uniformly satisfies the small-error proximal error-bound (EB) condition, we propose a first-order algorithm that converges to an $\epsilon$-stationary point of the penalty function, using in total $O(\epsilon^{-3})$ and $O(\epsilon^{-7})$ accesses to first-order (stochastic) gradient oracles when the oracle is deterministic and oracles are noisy, respectively.
no code implementations • 15 Jun 2023 • Haoyue Bai, Gregory Canal, Xuefeng Du, Jeongyeol Kwon, Robert Nowak, Yixuan Li
Modern machine learning models deployed in the wild can encounter both covariate and semantic shifts, giving rise to the problems of out-of-distribution (OOD) generalization and OOD detection respectively.
1 code implementation • NeurIPS 2023 • Jifan Zhang, Shuai Shao, Saurabh Verma, Robert Nowak
To address this, we propose the first adaptive algorithm selection strategy for deep active learning.
no code implementations • 29 Jan 2023 • Subhojyoti Mukherjee, Qiaomin Xie, Josiah Hanna, Robert Nowak
In this paper, we study the problem of optimal data collection for policy evaluation in linear bandits.
no code implementations • 26 Jan 2023 • Jeongyeol Kwon, Dohyun Kwon, Stephen Wright, Robert Nowak
Specifically, we show that F2SA converges to an $\epsilon$-stationary solution of the bilevel problem after $\epsilon^{-7/2}, \epsilon^{-5/2}$, and $\epsilon^{-3/2}$ iterations (each iteration using $O(1)$ samples) when stochastic noises are in both level objectives, only in the upper-level objective, and not present (deterministic settings), respectively.
no code implementations • 15 Oct 2022 • Yinglun Zhu, Robert Nowak
Deep neural networks have great representation power, but typically require large numbers of training examples.
no code implementations • 13 Oct 2022 • Przemysław Stawczyk, Robert Nowak
The algorithm consists of several steps, of which the most important are : (1) conversion of the restriction maps into binary strings, (2) detection of overlaps between restriction maps, (3) determining the layout of restriction maps set, (4) creation of consensus genomic maps.
2 code implementations • 7 Jul 2022 • Gregory Canal, Blake Mason, Ramya Korlakai Vinayak, Robert Nowak
This paper investigates simultaneous preference and metric learning from a crowd of respondents.
no code implementations • 31 Mar 2022 • Yinglun Zhu, Robert Nowak
Furthermore, our algorithm is guaranteed to only abstain on hard examples (where the true label distribution is close to a fair coin), a novel property we term \emph{proper abstention} that also leads to a host of other desirable characteristics (e. g., recovering minimax guarantees in the standard setting, and avoiding the undesirable ``noise-seeking'' behavior often seen in active learning).
no code implementations • 9 Mar 2022 • Subhojyoti Mukherjee, Josiah P. Hanna, Robert Nowak
This paper studies the problem of data collection for policy evaluation in Markov decision processes (MDPs).
1 code implementation • 7 Feb 2022 • Julian Katz-Samuels, Julia Nakhleh, Robert Nowak, Yixuan Li
Out-of-distribution (OOD) detection is important for machine learning models deployed in the wild.
BIG-bench Machine Learning Out of Distribution (OOD) Detection
1 code implementation • 3 Feb 2022 • Jifan Zhang, Julian Katz-Samuels, Robert Nowak
Active learning is a label-efficient approach to train highly effective models while interactively selecting only small subsets of unlabelled data for labelling and training.
no code implementations • 2 Nov 2021 • Blake Mason, Romain Camilleri, Subhojyoti Mukherjee, Kevin Jamieson, Robert Nowak, Lalit Jain
The threshold value $\alpha$ can either be \emph{explicit} and provided a priori, or \emph{implicit} and defined relative to the optimal function value, i. e. $\alpha = (1-\epsilon)f(x_\ast)$ for a given $\epsilon > 0$ where $f(x_\ast)$ is the maximal function value and is unknown.
no code implementations • 10 Sep 2021 • Yinglun Zhu, Julian Katz-Samuels, Robert Nowak
The core of our algorithms is a new optimization problem based on experimental design that leverages the geometry of the action set to identify a near-optimal hypothesis class.
no code implementations • NeurIPS 2021 • Yinglun Zhu, Dongruo Zhou, Ruoxi Jiang, Quanquan Gu, Rebecca Willett, Robert Nowak
To overcome the curse of dimensionality, we propose to adaptively embed the feature representation of each arm into a lower-dimensional space and carefully deal with the induced model misspecification.
no code implementations • 8 Mar 2021 • Blake Mason, Ardhendu Tripathy, Robert Nowak
Specifically, consider the setting in which an NNS algorithm has access only to a stochastic distance oracle that provides a noisy, unbiased estimate of the distance between any pair of points, rather than the exact distance.
no code implementations • 12 Feb 2021 • Yinglun Zhu, Robert Nowak
In this paper, we establish the first lower bound for the model selection problem.
no code implementations • 15 Dec 2020 • Subhojyoti Mukherjee, Ardhendu Tripathy, Robert Nowak
Active learning can reduce the number of samples needed to perform a hypothesis test and to estimate the parameters of a model.
no code implementations • NeurIPS 2020 • Blake Mason, Lalit Jain, Ardhendu Tripathy, Robert Nowak
The pure-exploration problem in stochastic multi-armed bandits aims to find one or more arms with the largest (or near largest) means.
1 code implementation • ICML 2020 • Yinglun Zhu, Sumeet Katariya, Robert Nowak
We study the problem of Robust Outlier Arm Identification (ROAI), where the goal is to identify arms whose expected rewards deviate substantially from the majority, by adaptively sampling from their reward distributions.
1 code implementation • 30 Jun 2020 • Cody Coleman, Edward Chou, Julian Katz-Samuels, Sean Culatana, Peter Bailis, Alexander C. Berg, Robert Nowak, Roshan Sumbaly, Matei Zaharia, I. Zeki Yalniz
Many active learning and search approaches are intractable for large-scale industrial settings with billions of unlabeled examples.
no code implementations • NeurIPS 2020 • Yinglun Zhu, Robert Nowak
With additional knowledge of the expected reward of the best arm, we propose another adaptive algorithm that is minimax optimal, up to polylog factors, over all hardness levels.
1 code implementation • 16 Jun 2020 • Blake Mason, Lalit Jain, Ardhendu Tripathy, Robert Nowak
Mathematically, the all-{\epsilon}-good arm identification problem presents significant new challenges and surprises that do not arise in the pure-exploration objectives studied in the past.
no code implementations • 6 Jun 2019 • Ayon Sen, Xiaojin Zhu, Liam Marshall, Robert Nowak
Adversarial attacks aim to confound machine learning systems, while remaining virtually imperceptible to humans.
1 code implementation • NeurIPS 2019 • Sumeet Katariya, Ardhendu Tripathy, Robert Nowak
This paper studies the problem of adaptively sampling from K distributions (arms) in order to identify the largest gap between any two adjacent means.
1 code implementation • NeurIPS 2019 • Blake Mason, Ardhendu Tripathy, Robert Nowak
We consider the problem of learning the nearest neighbor graph of a dataset of n items.
no code implementations • 9 Mar 2019 • Urvashi Oswal, Aniruddha Bhargava, Robert Nowak
In comparison, the regret of traditional linear bandits is $d\sqrt{T}$, where $d$ is the total number of (relevant and irrelevant) features, so the improvement can be dramatic if $k\ll d$.
no code implementations • 8 Jan 2019 • Kwang-Sung Jun, Rebecca Willett, Stephen Wright, Robert Nowak
We introduce the bilinear bandit problem with low-rank structure in which an action takes the form of a pair of arms from two different entity types, and the reward is a bilinear function of the known feature vectors of the arms.
no code implementations • 10 Jul 2018 • Urvashi Oswal, Robert Nowak
The main contribution of the paper is a new approach to subspace clustering that is significantly more computationally efficient and scalable than existing state-of-the-art methods.
no code implementations • 25 Feb 2018 • Yuzhe Ma, Robert Nowak, Philippe Rigollet, Xuezhou Zhang, Xiaojin Zhu
We call a learner super-teachable if a teacher can trim down an iid training set while making the learner learn even better.
1 code implementation • 20 Feb 2018 • Sumeet Katariya, Lalit Jain, Nandana Sengupta, James Evans, Robert Nowak
We consider the problem of active coarse ranking, where the goal is to sort items according to their means into clusters of pre-specified sizes, by adaptively sampling from their reward distributions.
no code implementations • NeurIPS 2017 • Ervin Tanczos, Robert Nowak, Bob Mankoff
This paper focuses on best-arm identification in multi-armed bandits with bounded rewards.
1 code implementation • AISTATS, Electronic Journal of Statistics 2017 • Daniel Pimentel-Alarcon, Robert Nowak
This paper presents r2pca, a random con- sensus method for robust principal compo- nent analysis.
no code implementations • NeurIPS 2017 • Lalit Jain, Blake Mason, Robert Nowak
This paper investigates the theoretical foundations of metric learning, focused on three key questions that are not fully addressed in prior work: 1) we consider learning general low-dimensional (low-rank) metrics as well as sparse metrics; 2) we develop upper and lower (minimax)bounds on the generalization error; 3) we quantify the sample complexity of metric learning in terms of the dimension of the feature space and the dimension/rank of the underlying metric;4) we also bound the accuracy of the learned metric relative to the underlying true generative metric.
no code implementations • 13 Jul 2017 • Gautam Dasarathy, Elchanan Mossel, Robert Nowak, Sebastien Roch
As a corollary, we also obtain a new identifiability result of independent interest: for any species tree with $n \geq 3$ species, the rooted species tree can be identified from the distribution of its unrooted weighted gene trees even in the absence of a molecular clock.
no code implementations • NeurIPS 2017 • Kwang-Sung Jun, Aniruddha Bhargava, Robert Nowak, Rebecca Willett
Second, for the case where the number $N$ of arms is very large, we propose new algorithms in which each next arm is selected via an inner product search.
no code implementations • 3 Sep 2016 • Kwang-Sung Jun, Robert Nowak
In graph-based active learning, algorithms based on expected error minimization (EEM) have been popular and yield good empirical performance.
no code implementations • NeurIPS 2016 • Lalit Jain, Kevin Jamieson, Robert Nowak
First, we derive prediction error bounds for ordinal embedding with noise by exploiting the fact that the rank of a distance matrix of points in $\mathbb{R}^d$ is at most $d+2$.
no code implementations • 14 Mar 2016 • Aniruddha Bhargava, Ravi Ganti, Robert Nowak
In this paper we model the problem of learning preferences of a population as an active learning problem.
no code implementations • 13 Mar 2016 • Nikhil Rao, Ravi Ganti, Laura Balzano, Rebecca Willett, Robert Nowak
Single Index Models (SIMs) are simple yet flexible semi-parametric models for machine learning, where the response variable is modeled as a monotonic function of a linear combination of features.
no code implementations • 30 Jun 2015 • Ravi Ganti, Nikhil Rao, Rebecca M. Willett, Robert Nowak
Single Index Models (SIMs) are simple yet flexible semi-parametric models for classification and regression.
no code implementations • 29 Jun 2015 • Gautam Dasarathy, Robert Nowak, Xiaojin Zhu
This paper investigates the problem of active learning for binary label prediction on a graph.
no code implementations • 31 Jan 2015 • Kevin Jamieson, Sumeet Katariya, Atul Deshpande, Robert Nowak
We prove that in the absence of structural assumptions, the sample complexity of this problem is proportional to the sum of the inverse squared gaps between the Borda scores of each suboptimal arm and the best arm.
no code implementations • 28 Apr 2014 • Gautam Dasarathy, Robert Nowak, Sebastien Roch
We consider the problem of estimating the evolutionary history of a set of species (phylogeny or species tree) from several genes.
no code implementations • 18 Feb 2014 • Nikhil Rao, Robert Nowak, Christopher Cox, Timothy Rogers
In this paper, we are interested in a less restrictive form of structured sparse feature selection: we assume that while features can be grouped according to some notion of similarity, not all features in a group need be selected for the task at hand.
no code implementations • 27 Dec 2013 • Kevin Jamieson, Matthew Malloy, Robert Nowak, Sébastien Bubeck
The paper proposes a novel upper confidence bound (UCB) procedure for identifying the arm with the largest mean in a multi-armed bandit game in the fixed confidence setting using a small number of total samples.
no code implementations • NeurIPS 2013 • Nikhil Rao, Christopher Cox, Robert Nowak, Timothy Rogers
In this paper, we are interested in a less restrictive form of multitask learning, wherein (1) the available features can be organized into subsets according to a notion of similarity and (2) features useful in one task are similar, but not necessarily identical, to the features best suited for other tasks.
no code implementations • 17 Jun 2013 • Kevin Jamieson, Matthew Malloy, Robert Nowak, Sebastien Bubeck
Motivated by large-scale applications, we are especially interested in identifying situations where the total number of samples that are necessary and sufficient to find the best arm scale linearly with the number of arms.
no code implementations • 17 Apr 2013 • Divyanshu Vats, Robert Nowak
We highlight three main properties of using junction trees for UGMS.
no code implementations • NeurIPS 2012 • Kevin G. Jamieson, Robert Nowak, Ben Recht
Moreover, if the function evaluations are noisy, then approximating gradients by finite differences is difficult.
no code implementations • NeurIPS 2010 • Andrew Goldberg, Ben Recht, Jun-Ming Xu, Robert Nowak, Jerry Zhu
We pose transductive classification as a matrix completion problem.
1 code implementation • 21 Jun 2010 • Laura Balzano, Robert Nowak, Benjamin Recht
GROUSE performs exceptionally well in practice both in tracking subspaces and as an online algorithm for matrix completion.
no code implementations • NeurIPS 2009 • Robert Nowak
This paper addresses the problem of noisy Generalized Binary Search (GBS).
no code implementations • NeurIPS 2008 • Aarti Singh, Robert Nowak, Jerry Zhu
We show that there are large classes of problems for which SSL can significantly outperform supervised learning, in finite sample regimes and sometimes also in terms of error convergence rates.
no code implementations • NeurIPS 2008 • Rui M. Castro, Charles Kalish, Robert Nowak, Ruichen Qian, Tim Rogers, Jerry Zhu
We investigate a topic at the interface of machine learning and cognitive science.