no code implementations • 2 Apr 2024 • Yu Xia, Xu Liu, Tong Yu, Sungchul Kim, Ryan A. Rossi, Anup Rao, Tung Mai, Shuai Li
Large Language Models (LLMs) have shown propensity to generate hallucinated outputs, i. e., texts that are factually incorrect or unsupported.
no code implementations • 5 Apr 2023 • Tung Mai, Alexander Munteanu, Cameron Musco, Anup B. Rao, Chris Schwiegelshohn, David P. Woodruff
For this problem, under the $\ell_2$ norm, we observe an upper bound of $O(k \log (d)/\varepsilon + k\log(k/\varepsilon)/\varepsilon^2)$ rows, showing that sparse recovery is strictly easier to sketch than sparse regression.
1 code implementation • 12 Oct 2022 • Raghavendra Addanki, David Arbour, Tung Mai, Cameron Musco, Anup Rao
In particular, we study sample-constrained treatment effect estimation, where we must select a subset of $s \ll n$ individuals from the population to experiment on.
no code implementations • 28 Jan 2022 • Nikhil Sheoran, Subrata Mitra, Vibhor Porwal, Siddharth Ghetia, Jatin Varshney, Tung Mai, Anup Rao, Vikas Maddukuri
The goal of Approximate Query Processing (AQP) is to provide very fast but "accurate enough" results for costly aggregate queries thereby improving user experience in interactive exploration of large datasets.
no code implementations • 29 Nov 2021 • Aravind Reddy, Ryan A. Rossi, Zhao Song, Anup Rao, Tung Mai, Nedim Lipka, Gang Wu, Eunyee Koh, Nesreen Ahmed
In this paper, we introduce the online and streaming MAP inference and learning problems for Non-symmetric Determinantal Point Processes (NDPPs) where data points arrive in an arbitrary order and the algorithms are constrained to use a single-pass over the data as well as sub-linear memory.
no code implementations • NeurIPS 2021 • Tung Mai, Anup B. Rao, Cameron Musco
It also does not depend on the specific loss function, so a single coreset can be used in multiple training scenarios.
no code implementations • NeurIPS 2021 • Tung Mai, Cameron N Musco, Anup Rao
It also does not depend on the specific loss function, so a single coreset can be used in multiple training scenarios.
no code implementations • 8 Mar 2021 • Mojtaba Sahraee-Ardakan, Tung Mai, Anup Rao, Ryan Rossi, Sundeep Rangan, Alyson K. Fletcher
We show the double descent phenomenon in our experiments for convolutional models and show that our theoretical results match the experiments.
no code implementations • 25 Feb 2021 • Enayat Ullah, Tung Mai, Anup Rao, Ryan Rossi, Raman Arora
Our key contribution is the design of corresponding efficient unlearning algorithms, which are based on constructing a (maximal) coupling of Markov chains for the noisy SGD procedure.
no code implementations • 4 Feb 2021 • David Arbour, Drew Dimmery, Tung Mai, Anup Rao
We study the online discrepancy minimization problem for vectors in $\mathbb{R}^d$ in the oblivious setting where an adversary is allowed fix the vectors $x_1, x_2, \ldots, x_n$ in arbitrary order ahead of time.
Data Structures and Algorithms Discrete Mathematics Combinatorics
no code implementations • 15 Jan 2021 • Mohammad Mehrabi, Adel Javanmard, Ryan A. Rossi, Anup Rao, Tung Mai
We study the tradeoff between standard risk and adversarial risk and derive the Pareto-optimal tradeoff, achievable over specific classes of models, in the infinite data limit with features dimension kept fixed.
1 code implementation • 28 Sep 2020 • Jiong Zhu, Ryan A. Rossi, Anup Rao, Tung Mai, Nedim Lipka, Nesreen K. Ahmed, Danai Koutra
Graph Neural Networks (GNNs) have proven to be useful for many different practical applications.