no code implementations • 23 Apr 2024 • Avi Schwarzschild, Zhili Feng, Pratyush Maini, Zachary C. Lipton, J. Zico Kolter
The ACR overcomes the limitations of existing notions of memorization by (i) offering an adversarial view of measuring memorization, especially for monitoring unlearning and compliance; and (ii) allowing for the flexibility to measure memorization for arbitrary strings at a reasonably low compute.
no code implementations • 15 Apr 2024 • Yiming Zhang, Zhuokai Zhao, Zhaorun Chen, Zhili Feng, Zenghui Ding, Yining Sun
Self-supervised contrastive learning models, such as CLIP, have set new benchmarks for vision-language models in many downstream tasks.
no code implementations • 12 Jan 2024 • Zhili Feng, Michal Moshkovitz, Dotan Di Castro, J. Zico Kolter
Concept explanation is a popular approach for examining how human-interpretable concepts impact the predictions of a model.
2 code implementations • 11 Jan 2024 • Pratyush Maini, Zhili Feng, Avi Schwarzschild, Zachary C. Lipton, J. Zico Kolter
Large language models trained on massive corpora of data from the web can memorize and reproduce sensitive or private data raising both legal and ethical concerns.
no code implementations • 21 Oct 2023 • Zhili Feng, J. Zico Kolter
This work studies the neural tangent kernel (NTK) of the deep equilibrium (DEQ) model, a practical ``infinite-depth'' architecture which directly computes the infinite-depth limit of a weight-tied network via root-finding.
no code implementations • 11 Jul 2023 • Zhili Feng, Ezra Winston, J. Zico Kolter
Deep Boltzmann machines (DBMs), one of the first ``deep'' learning methods ever studied, are multi-layered probabilistic models governed by a pairwise energy function that describes the likelihood of all variables/nodes in the network.
no code implementations • 10 Jul 2023 • Zhili Feng, Anna Bair, J. Zico Kolter
This method first automatically generates multiple visual descriptions of each class via a large language model (LLM), then uses a VLM to translate these descriptions to a set of visual feature embeddings of each image, and finally uses sparse logistic regression to select a relevant subset of these features to classify each image.
no code implementations • ICLR 2022 • Jon C. Ergun, Zhili Feng, Sandeep Silwal, David P. Woodruff, Samson Zhou
$k$-means clustering is a well-studied problem due to its wide applicability.
no code implementations • 16 Jun 2021 • Zhili Feng, Fred Roosta, David P. Woodruff
In this paper, we present novel dimensionality reduction methods for non-PSD matrices, as well as their ``square-roots", which involve matrices with complex entries.
no code implementations • ICLR 2022 • Zhili Feng, Shaobo Han, Simon S. Du
This paper studies zero-shot domain adaptation where each domain is indexed on a multi-dimensional array, and we only have data from a small subset of domains.
no code implementations • 12 Jun 2019 • Qiang Ning, Ben Zhou, Zhili Feng, Haoruo Peng, Dan Roth
Automatic extraction of temporal information in text is an important component of natural language understanding.
no code implementations • EMNLP 2017 • Qiang Ning, Zhili Feng, Dan Roth
Identifying temporal relations between events is an essential step towards natural language understanding.
Ranked #1 on Temporal Information Extraction on TempEval-3
no code implementations • ACL 2018 • Qiang Ning, Zhili Feng, Hao Wu, Dan Roth
Understanding temporal and causal relations between events is a fundamental natural language understanding task.
no code implementations • 8 May 2019 • Shashank Rajput, Zhili Feng, Zachary Charles, Po-Ling Loh, Dimitris Papailiopoulos
Data augmentation (DA) is commonly used during model training, as it significantly improves test error and model robustness.
no code implementations • EMNLP 2018 • Qiang Ning, Ben Zhou, Zhili Feng, Haoruo Peng, Dan Roth
Automatic extraction of temporal information is important for natural language understanding.
1 code implementation • LREC 2018 • Daniel Khashabi, Mark Sammons, Ben Zhou, Tom Redman, Christos Christodoulopoulos, Vivek Srikumar, Nicholas Rizzolo, Lev Ratinov, Guanheng Luo, Quang Do, Chen-Tse Tsai, Subhro Roy, Stephen Mayhew, Zhili Feng, John Wieting, Xiaodong Yu, Yangqiu Song, Shashank Gupta, Shyam Upadhyay, Naveen Arivazhagan, Qiang Ning, Shaoshi Ling, Dan Roth
no code implementations • 1 Apr 2018 • Zhili Feng, Po-Ling Loh
When the adversary is allowed a bounded memory of size 1, we show that a matching lower bound of $\widetilde\Omega(T^{2/3})$ is achieved in the case of full-information feedback.