Search Results for author: Tung Mai

Found 12 papers, 2 papers with code

Hallucination Diversity-Aware Active Learning for Text Summarization

no code implementations2 Apr 2024 Yu Xia, Xu Liu, Tong Yu, Sungchul Kim, Ryan A. Rossi, Anup Rao, Tung Mai, Shuai Li

Large Language Models (LLMs) have shown propensity to generate hallucinated outputs, i. e., texts that are factually incorrect or unsupported.

Active Learning Hallucination +1

Optimal Sketching Bounds for Sparse Linear Regression

no code implementations5 Apr 2023 Tung Mai, Alexander Munteanu, Cameron Musco, Anup B. Rao, Chris Schwiegelshohn, David P. Woodruff

For this problem, under the $\ell_2$ norm, we observe an upper bound of $O(k \log (d)/\varepsilon + k\log(k/\varepsilon)/\varepsilon^2)$ rows, showing that sparse recovery is strictly easier to sketch than sparse regression.

regression

Sample Constrained Treatment Effect Estimation

1 code implementation12 Oct 2022 Raghavendra Addanki, David Arbour, Tung Mai, Cameron Musco, Anup Rao

In particular, we study sample-constrained treatment effect estimation, where we must select a subset of $s \ll n$ individuals from the population to experiment on.

Causal Inference

Electra: Conditional Generative Model based Predicate-Aware Query Approximation

no code implementations28 Jan 2022 Nikhil Sheoran, Subrata Mitra, Vibhor Porwal, Siddharth Ghetia, Jatin Varshney, Tung Mai, Anup Rao, Vikas Maddukuri

The goal of Approximate Query Processing (AQP) is to provide very fast but "accurate enough" results for costly aggregate queries thereby improving user experience in interactive exploration of large datasets.

Online MAP Inference and Learning for Nonsymmetric Determinantal Point Processes

no code implementations29 Nov 2021 Aravind Reddy, Ryan A. Rossi, Zhao Song, Anup Rao, Tung Mai, Nedim Lipka, Gang Wu, Eunyee Koh, Nesreen Ahmed

In this paper, we introduce the online and streaming MAP inference and learning problems for Non-symmetric Determinantal Point Processes (NDPPs) where data points arrive in an arbitrary order and the algorithms are constrained to use a single-pass over the data as well as sub-linear memory.

Point Processes valid

Coresets for Classification -- Simplified and Strengthened

no code implementations NeurIPS 2021 Tung Mai, Anup B. Rao, Cameron Musco

It also does not depend on the specific loss function, so a single coreset can be used in multiple training scenarios.

Active Learning Classification

Coresets for Classification – Simplified and Strengthened

no code implementations NeurIPS 2021 Tung Mai, Cameron N Musco, Anup Rao

It also does not depend on the specific loss function, so a single coreset can be used in multiple training scenarios.

Active Learning Classification

Asymptotics of Ridge Regression in Convolutional Models

no code implementations8 Mar 2021 Mojtaba Sahraee-Ardakan, Tung Mai, Anup Rao, Ryan Rossi, Sundeep Rangan, Alyson K. Fletcher

We show the double descent phenomenon in our experiments for convolutional models and show that our theoretical results match the experiments.

regression

Machine Unlearning via Algorithmic Stability

no code implementations25 Feb 2021 Enayat Ullah, Tung Mai, Anup Rao, Ryan Rossi, Raman Arora

Our key contribution is the design of corresponding efficient unlearning algorithms, which are based on constructing a (maximal) coupling of Markov chains for the noisy SGD procedure.

Machine Unlearning

Online Discrepancy Minimization via Persistent Self-Balancing Walks

no code implementations4 Feb 2021 David Arbour, Drew Dimmery, Tung Mai, Anup Rao

We study the online discrepancy minimization problem for vectors in $\mathbb{R}^d$ in the oblivious setting where an adversary is allowed fix the vectors $x_1, x_2, \ldots, x_n$ in arbitrary order ahead of time.

Data Structures and Algorithms Discrete Mathematics Combinatorics

Fundamental Tradeoffs in Distributionally Adversarial Training

no code implementations15 Jan 2021 Mohammad Mehrabi, Adel Javanmard, Ryan A. Rossi, Anup Rao, Tung Mai

We study the tradeoff between standard risk and adversarial risk and derive the Pareto-optimal tradeoff, achievable over specific classes of models, in the infinite data limit with features dimension kept fixed.

Binary Classification regression

Graph Neural Networks with Heterophily

1 code implementation28 Sep 2020 Jiong Zhu, Ryan A. Rossi, Anup Rao, Tung Mai, Nedim Lipka, Nesreen K. Ahmed, Danai Koutra

Graph Neural Networks (GNNs) have proven to be useful for many different practical applications.

Cannot find the paper you are looking for? You can Submit a new open access paper.