no code implementations • NAACL 2022 • Cynthia Sullivan, William Brackenbury, Andrew McNut, Kevin Bryson, Kbyllofficial@gmail.com Kbyllofficial@gmail.com, Yuxin Chen, Michael Littman, Chenhao Tan, Blase Ur
In the context of data labeling, NLP researchers are increasingly interested in having humans select rationales, a subset of input tokens relevant to the chosen label.
no code implementations • 22 Jan 2025 • Jiadong Liang, Zhihan Huang, Yuxin Chen
This paper investigates how diffusion generative models leverage (unknown) low-dimensional structure to accelerate sampling.
1 code implementation • 16 Jan 2025 • Chaoqi Wang, Zhuokai Zhao, Yibo Jiang, Zhaorun Chen, Chen Zhu, Yuxin Chen, Jiayi Liu, Lizhu Zhang, Xiangjun Fan, Hao Ma, Sinong Wang
As a drop-in enhancement to the existing RLHF workflow, our causal reward modeling provides a practical way to improve the trustworthiness and fairness of LLM finetuning.
no code implementations • 14 Jan 2025 • Yuxin Chen, Devesh K. Jha, Masayoshi Tomizuka, Diego Romeres
This reward is then used to fine-tune the pre-trained policy with reinforcement learning (RL), resulting in alignment of pre-trained policy with new human preferences while still solving the original task.
no code implementations • 13 Jan 2025 • Reza Jalayer, Yuxin Chen, Masoud Jalayer, Carlotta Orsenigo, Masayoshi Tomizuka
This approach allowed us to account for multiple camera perspectives while also evaluating the performance of models trained on existing egocentric datasets as well as static-camera datasets.
no code implementations • 26 Nov 2024 • Zihan Zhang, Jason D. Lee, Simon S. Du, Yuxin Chen
This work investigates stepsize-based acceleration of gradient descent with {\em anytime} convergence guarantees.
no code implementations • 26 Nov 2024 • Yinan Zhou, Yuxin Chen, Haokun Lin, Shuyu Yang, Li Zhu, Zhongang Qi, Chen Ma, Ying Shan
In recent years, Multimodal Large Language Models (MLLMs) have increasingly emphasized grounding and referring capabilities to achieve detailed understanding and flexible user interaction.
no code implementations • 22 Nov 2024 • Tao Zhang, Ziqi Zhang, Zongyang Ma, Yuxin Chen, Zhongang Qi, Chunfeng Yuan, Bing Li, Junfu Pu, Yuxuan Zhao, Zehua Xie, Jin Ma, Ying Shan, Weiming Hu
Thus, multimodal Retrieval-Augmented Generation (mRAG) is naturally introduced to provide MLLMs with comprehensive and up-to-date knowledge, effectively expanding the knowledge scope.
no code implementations • 15 Nov 2024 • Zhichen Zeng, Xiaolong Liu, Mengyue Hang, Xiaoyi Liu, Qinghai Zhou, Chaofei Yang, Yiqun Liu, Yichen Ruan, Laming Chen, Yuxin Chen, Yujia Hao, Jiaqi Xu, Jade Nie, Xi Liu, Buyun Zhang, Wei Wen, Siyang Yuan, Kai Wang, Wen-Yen Chen, Yiping Han, Huayu Li, Chunzhi Yang, Bo Long, Philip S. Yu, Hanghang Tong, Jiyan Yang
A mutually beneficial integration of heterogeneous information is the cornerstone towards the success of CTR prediction.
1 code implementation • 7 Nov 2024 • Jiangshan Wang, Junfu Pu, Zhongang Qi, Jiayi Guo, Yue Ma, Nisha Huang, Yuxin Chen, Xiu Li, Ying Shan
To address this issue, we propose RF-Solver, a novel training-free sampler that effectively enhances inversion precision by mitigating the errors in the ODE-solving process of rectified flow.
no code implementations • 6 Nov 2024 • Diantong Li, Fengxue Zhang, Chong Liu, Yuxin Chen
Multi-objective Bayesian optimization has been widely adopted in scientific experiment design, including drug discovery and hyperparameter optimization.
no code implementations • 30 Oct 2024 • Shuzhen Li, Yuxin Chen, Xuesong Chen, Ruiyang Gao, Yupeng Zhang, Chao Yu, Yunfei Li, Ziyi Ye, Weijun Huang, Hongliang Yi, Yue Leng, Yi Wu
However, reliable BCG-based sleep staging is challenging due to the limited sleep monitoring data available for BCG.
no code implementations • 25 Oct 2024 • Yuxin Chen, Zijian Wu, Adam Schmidt, Septimiu E. Salcudean
Methods: We use the Segment Anything Model2 (SAM2) to detect and mask occlusions by surgical tools, and we develop and integrate into SENDD an Adaptive Multi-Flow Sparse Tracker (A-MFST) with forward-backward consistency metrics, to enhance occlusion and uncertainty estimation.
no code implementations • 24 Oct 2024 • Zhihan Huang, Yuting Wei, Yuxin Chen
The denoising diffusion probabilistic model (DDPM) has emerged as a mainstream generative model in generative AI.
no code implementations • 16 Oct 2024 • Chaoqi Wang, Zhuokai Zhao, Chen Zhu, Karthik Abinav Sankararaman, Michal Valko, Xuefei Cao, Zhaorun Chen, Madian Khabsa, Yuxin Chen, Hao Ma, Sinong Wang
However, current post-training methods such as reinforcement learning from human feedback (RLHF) and direct alignment from preference methods (DAP) primarily utilize single-sample comparisons.
no code implementations • 7 Oct 2024 • Yuchen Wu, Yuxin Chen, Yuting Wei
Diffusion models play a pivotal role in contemporary generative modeling, claiming state-of-the-art performance across various domains.
1 code implementation • 20 Sep 2024 • Yu Zhang, Changhao Pan, Wenxiang Guo, RuiQi Li, Zhiyuan Zhu, Jialei Wang, Wenhao Xu, Jingyu Lu, Zhiqing Hong, Chuxin Wang, Lichao Zhang, Jinzheng He, Ziyue Jiang, Yuxin Chen, Chen Yang, Jiecheng Zhou, Xinyu Cheng, Zhou Zhao
The scarcity of high-quality and multi-task singing datasets significantly hinders the development of diverse controllable and personalized singing tasks, as existing singing datasets suffer from low quality, limited diversity of languages and singers, absence of multi-technique information and realistic music scores, and poor task suitability.
no code implementations • 5 Aug 2024 • Gen Li, Yuting Wei, Yuejie Chi, Yuxin Chen
Diffusion models, which convert noise into new data instances by learning to reverse a diffusion process, have become a cornerstone in contemporary generative modeling.
no code implementations • 10 Jul 2024 • Zongyang Ma, Ziqi Zhang, Yuxin Chen, Zhongang Qi, Chunfeng Yuan, Bing Li, Yingmin Luo, Xu Li, Xiaojuan Qi, Ying Shan, Weiming Hu
EA-VTR can efficiently encode frame-level and video-level visual representations simultaneously, enabling detailed event content and complex event temporal cross-modal alignment, ultimately enhancing the comprehensive understanding of video events.
no code implementations • CVPR 2024 • Yuxin Chen, Zongyang Ma, Ziqi Zhang, Zhongang Qi, Chunfeng Yuan, Bing Li, Junfu Pu, Ying Shan, Xiaojuan Qi, Weiming Hu
Dominant dual-encoder models enable efficient image-text retrieval but suffer from limited accuracy while the cross-encoder models offer higher accuracy at the expense of efficiency.
1 code implementation • 7 Jul 2024 • Leheng Sheng, An Zhang, Yi Zhang, Yuxin Chen, Xiang Wang, Tat-Seng Chua
Contrary to prevailing understanding that LMs and traditional recommenders learn two distinct representation spaces due to the huge gap in language and behavior modeling objectives, this work re-examines such understanding and explores extracting a recommendation space directly from the language representation space.
no code implementations • 5 Jul 2024 • Rui Ding, Jianguo Liu, Kang Hua, Xuebin Wang, Xiaoben Zhang, Minhua Shao, Yuxin Chen, Junhong Chen
The process culminated in the discovery of a promising Ru-Mn-Ca-Pr oxide catalyst.
1 code implementation • 27 Jun 2024 • Yicheng Xu, Yuxin Chen, Jiahao Nie, Yusong Wang, Huiping Zhuang, Manabu Okumura
In this setting, a CL learner is required to incrementally learn from multiple domains and classify test images from both seen and unseen domains without any domain-identity hint.
no code implementations • 24 Jun 2024 • Yuxin Chen, Chen Tang, Chenran Li, Ran Tian, Wei Zhan, Peter Stone, Masayoshi Tomizuka
Instead of inferring the complete human behavior characteristics, MEReQ infers a residual reward function that captures the discrepancy between the human expert's and the prior policy's underlying reward functions.
1 code implementation • 13 Jun 2024 • Yuxin Chen, Junfei Tan, An Zhang, Zhengyi Yang, Leheng Sheng, Enzhi Zhang, Xiang Wang, Tat-Seng Chua
Specifically, we incorporate multiple negatives in user preference data and devise an alternative version of DPO loss tailored for LM-based recommenders, which is extended from the traditional full-ranking Plackett-Luce (PL) model to partial rankings and connected to softmax sampling strategies.
no code implementations • 14 May 2024 • Minbiao Han, Fengxue Zhang, Yuxin Chen
This paper investigates the challenge of learning in black-box games, where the underlying utility function is unknown to any of the agents.
no code implementations • 15 Mar 2024 • Zihan Zhang, Jason D. Lee, Yuxin Chen, Simon S. Du
A recent line of works showed regret bounds in reinforcement learning (RL) can be (nearly) independent of planning horizon, a. k. a.~the horizon-free bounds.
no code implementations • 6 Mar 2024 • Gen Li, Yu Huang, Timofey Efimov, Yuting Wei, Yuejie Chi, Yuxin Chen
Score-based diffusion models, while achieving remarkable empirical performance, often suffer from low sampling speed, due to extensive function evaluations needed during the sampling phase.
1 code implementation • 4 Mar 2024 • Buyun Zhang, Liang Luo, Yuxin Chen, Jade Nie, Xi Liu, Daifeng Guo, Yanli Zhao, Shen Li, Yuchen Hao, Yantao Yao, Guna Lakshminarayanan, Ellie Dingqiao Wen, Jongsoo Park, Maxim Naumov, Wenlin Chen
Scaling laws play an instrumental role in the sustainable improvement in model quality.
2 code implementations • 1 Mar 2024 • Liang Luo, Buyun Zhang, Michael Tsang, Yinbin Ma, Ching-Hsiang Chu, Yuxin Chen, Shen Li, Yuchen Hao, Yanli Zhao, Guna Lakshminarayanan, Ellie Dingqiao Wen, Jongsoo Park, Dheevatsa Mudigere, Maxim Naumov
We study a mismatch between the deep learning recommendation models' flat architecture, common distributed training paradigm and hierarchical data center topology.
no code implementations • 8 Feb 2024 • Zhuokai Zhao, Yibo Jiang, Yuxin Chen
Active Learning (AL) has gained prominence in integrating data-intensive machine learning (ML) models into domains with limited labeled data.
no code implementations • 11 Dec 2023 • Xiang Cheng, Yuxin Chen, Suvrit Sra
Many neural network architectures are known to be Turing Complete, and can thus, in principle implement arbitrary algorithms.
no code implementations • 8 Dec 2023 • Zihan Zhang, Wenhao Zhan, Yuxin Chen, Simon S. Du, Jason D. Lee
Focusing on a hypothesis class of Vapnik-Chervonenkis (VC) dimension d, we propose a novel algorithm that yields an varepsilon-optimal randomized hypothesis with a sample complexity on the order of (d+k)/varepsilon^2 (modulo some logarithmic factor), matching the best-known lower bound.
no code implementations • 14 Nov 2023 • Wei Wen, Kuang-Hung Liu, Igor Fedorov, Xin Zhang, Hang Yin, Weiwei Chu, Kaveh Hassani, Mengying Sun, Jiang Liu, Xu Wang, Lin Jiang, Yuxin Chen, Buyun Zhang, Xi Liu, Dehua Cheng, Zhengxing Chen, Guang Zhao, Fangqiu Han, Jiyan Yang, Yuchen Hao, Liang Xiong, Wen-Yen Chen
In industry system, such as ranking system in Meta, it is unclear whether NAS algorithms from the literature can outperform production baselines because of: (1) scale - Meta ranking systems serve billions of users, (2) strong baselines - the baselines are production models optimized by hundreds to thousands of world-class engineers for years since the rise of deep learning, (3) dynamic baselines - engineers may have established new and stronger baselines during NAS search, and (4) efficiency - the search pipeline must yield results quickly in alignment with the productionization life cycle.
no code implementations • 14 Nov 2023 • Hang Yin, Kuang-Hung Liu, Mengying Sun, Yuxin Chen, Buyun Zhang, Jiang Liu, Vivek Sehgal, Rudresh Rajnikant Panchal, Eugen Hotaj, Xi Liu, Daifeng Guo, Jamey Zhang, Zhou Wang, Shali Jiang, Huayu Li, Zhengxing Chen, Wen-Yen Chen, Jiyan Yang, Wei Wen
The large scale of models and tight production schedule requires AutoML to outperform human baselines by only using a small number of model evaluation trials (around 100).
no code implementations • 9 Nov 2023 • Renyu Zhang, Aly A. Khan, Yuxin Chen, Robert L. Grossman
Our experimental results demonstrate the effectiveness of our approach, showcasing improved classification performance compared to traditional single-instance label-based methods.
no code implementations • 4 Nov 2023 • Yuchen Zhou, Yuxin Chen
Tensor clustering, which seeks to extract underlying cluster structures from noisy tensor observations, has gained increasing attention.
no code implementations • 1 Nov 2023 • Tong Yang, Shicong Cen, Yuting Wei, Yuxin Chen, Yuejie Chi
Federated reinforcement learning (RL) enables collaborative decision making of multiple distributed agents without sharing local data trajectories.
no code implementations • 25 Oct 2023 • Zixin Ding, Si Chen, Ruoxi Jia, Yuxin Chen
To address these limitations, we propose a novel approach for active learning, which aims to select batches of unlabeled instances through a learned surrogate model for data acquisition.
1 code implementation • 16 Oct 2023 • An Zhang, Yuxin Chen, Leheng Sheng, Xiang Wang, Tat-Seng Chua
Recommender systems are the cornerstone of today's information dissemination, yet a disconnect between offline metrics and online performance greatly hinders their development.
no code implementations • 12 Oct 2023 • Fengxue Zhang, Zejie Zhu, Yuxin Chen
Optimizing objectives under constraints, where both the objectives and constraints are black box functions, is a common scenario in real-world applications such as scientific experimental design, design of medical therapies, and industrial process optimization.
no code implementations • 11 Oct 2023 • Yuxin Chen, Chen Tang, Ran Tian, Chenran Li, Jinning Li, Masayoshi Tomizuka, Wei Zhan
We observe that, generally, a more diverse set of co-play agents during training enhances the generalization performance of the ego agent; however, this improvement varies across distinct scenarios and environments.
no code implementations • 3 Oct 2023 • Xuefeng Liu, Takuma Yoneda, Rick L. Stevens, Matthew R. Walter, Yuxin Chen
Integral to RPI are Robust Active Policy Selection (RAPS) and Robust Policy Gradient (RPG), both of which reason over whether to perform state-wise imitation from the oracles or learn from its own value function when the learner's performance surpasses that of the oracles in a specific state.
1 code implementation • 28 Sep 2023 • Chaoqi Wang, Yibo Jiang, Chenghao Yang, Han Liu, Yuxin Chen
The increasing capabilities of large language models (LLMs) raise opportunities for artificial general intelligence but concurrently amplify safety concerns, such as potential misuse of AI systems, necessitating effective AI alignment.
no code implementations • 21 Sep 2023 • Luyao He, Zhongbao Zhang, Sen Su, Yuxin Chen
To address these issues, we propose BitCoin, an innovative Bidirectional tagging and supervised Contrastive learning based joint relational triple extraction framework.
no code implementations • 25 Jul 2023 • Zihan Zhang, Yuxin Chen, Jason D. Lee, Simon S. Du
While a number of recent works achieved asymptotically minimal regret in online RL, the optimality of these results is only guaranteed in a ``large-sample'' regime, imposing enormous burn-in cost in order for their algorithms to operate optimally.
no code implementations • 25 Jul 2023 • Fengxue Zhang, Jialin Song, James Bowden, Alexander Ladd, Yisong Yue, Thomas A. Desautels, Yuxin Chen
Our approach is easy to tune, and is able to focus on local region of the optimization space that can be tackled by existing BO methods.
1 code implementation • 17 Jun 2023 • Xuefeng Liu, Takuma Yoneda, Chaoqi Wang, Matthew R. Walter, Yuxin Chen
We introduce MAPS and MAPS-SE, a class of policy improvement algorithms that perform imitation learning from multiple suboptimal oracles.
no code implementations • 15 Jun 2023 • Gen Li, Yuting Wei, Yuxin Chen, Yuejie Chi
Diffusion models, which convert noise into new data instances by learning to reverse a Markov diffusion process, have become a cornerstone in contemporary generative modeling.
no code implementations • NeurIPS 2023 • Laixi Shi, Gen Li, Yuting Wei, Yuxin Chen, Matthieu Geist, Yuejie Chi
Assuming access to a generative model that draws samples based on the nominal MDP, we characterize the sample complexity of RMDPs when the uncertainty set is specified via either the total variation (TV) distance or $\chi^2$ divergence.
no code implementations • 3 May 2023 • Arman Rahbar, Ziyu Ye, Yuxin Chen, Morteza Haghir Chehreghani
Specifically, we employ a surrogate information acquisition function based on adaptive submodularity to actively query feature values with a minimal cost, while using a posterior sampling scheme to maintain a low regret for online prediction.
no code implementations • 14 Apr 2023 • Gen Li, Yuling Yan, Yuxin Chen, Jianqing Fan
This paper studies reward-agnostic exploration in reinforcement learning (RL) -- a scenario where the learner is unware of the reward functions during the exploration stage -- and designs an algorithm that improves over the state of the art.
no code implementations • 5 Apr 2023 • Yuxin Chen, Anna Goodridge, Manish Sahu, Aditi Kishore, Seena Vafaee, Harsha Mohan, Katherina Sapozhnikov, Francis Creighton, Russell Taylor, Deepa Galaiya
Results: The force measurements on the tip of the surgical drill are validated with raw-egg drilling experiments, where a force sensor mounted below the egg serves as ground truth.
no code implementations • 10 Mar 2023 • Yuchen Zhou, Yuxin Chen
This paper is concerned with estimating the column subspace of a low-rank matrix $\boldsymbol{X}^\star \in \mathbb{R}^{n_1\times n_2}$ from contaminated data.
1 code implementation • 6 Mar 2023 • Han Liu, Yizhou Tian, Chacha Chen, Shi Feng, Yuxin Chen, Chenhao Tan
Despite the promising performance of supervised learning, representations learned by supervised models may not align well with human intuitions: what models consider as similar examples can be perceived as distinct by humans.
no code implementations • 30 Jan 2023 • Gen Li, Yanxi Chen, Yu Huang, Yuejie Chi, H. Vincent Poor, Yuxin Chen
Efficient computation of the optimal transport distance between two distributions serves as an algorithm subroutine that empowers various applications.
no code implementations • ICCV 2023 • Zongyang Ma, Ziqi Zhang, Yuxin Chen, Zhongang Qi, Yingmin Luo, Zekun Li, Chunfeng Yuan, Bing Li, XiaoHu Qie, Ying Shan, Weiming Hu
This paper proposes a novel generative model, Order-Prompted Tag Sequence Generation (OP-TSG), according to the above characteristics.
no code implementations • CVPR 2023 • Yuxin Chen, Zongyang Ma, Ziqi Zhang, Zhongang Qi, Chunfeng Yuan, Ying Shan, Bing Li, Weiming Hu, XiaoHu Qie, Jianping Wu
ViLEM then enforces the model to discriminate the correctness of each word in the plausible negative texts and further correct the wrong words via resorting to image information.
Ranked #45 on Visual Reasoning on Winoground
no code implementations • 22 Aug 2022 • Gen Li, Yuejie Chi, Yuting Wei, Yuxin Chen
This paper studies multi-agent reinforcement learning in Markov games, with the goal of learning Nash equilibria or coarse correlated equilibria (CCE) sample-optimally.
no code implementations • 13 Jul 2022 • Xuefeng Liu, Fangfang Xia, Rick L. Stevens, Yuxin Chen
In particular, we focus on the task of selecting pre-trained classifiers, and propose a contextual active model selection algorithm (CAMS), which relies on a novel uncertainty sampling query criterion defined on a given policy class for adaptive model selection.
no code implementations • 6 Jul 2022 • Yifan Lu, Ziqi Zhang, Yuxin Chen, Chunfeng Yuan, Bing Li, Weiming Hu
The task of Dense Video Captioning (DVC) aims to generate captions with timestamps for multiple events in one video.
no code implementations • 8 Jun 2022 • Huiying Li, Arjun Nitin Bhagoji, Yuxin Chen, Haitao Zheng, Ben Y. Zhao
Existing research on training-time attacks for deep neural networks (DNNs), such as backdoors, largely assume that models are static once trained, and hidden backdoors trained into models remain active indefinitely.
no code implementations • 8 Jun 2022 • Yuling Yan, Gen Li, Yuxin Chen, Jianqing Fan
This paper makes progress towards learning Nash equilibria in two-player zero-sum Markov games from offline data.
Model-based Reinforcement Learning reinforcement-learning +1
no code implementations • 11 Apr 2022 • Gen Li, Laixi Shi, Yuxin Chen, Yuejie Chi, Yuting Wei
We demonstrate that the model-based (or "plug-in") approach achieves minimax-optimal sample complexity without burn-in cost for tabular Markov decision processes (MDPs).
no code implementations • 31 Mar 2022 • Ziqi Zhang, Yuxin Chen, Zongyang Ma, Zhongang Qi, Chunfeng Yuan, Bing Li, Ying Shan, Weiming Hu
In this paper, we propose to CREATE, the first large-scale Chinese shoRt vidEo retrievAl and Title gEneration benchmark, to facilitate research and application in video titling and video retrieval in Chinese.
1 code implementation • CVPR 2022 • Zongyang Ma, Guan Luo, Jin Gao, Liang Li, Yuxin Chen, Shaoru Wang, Congxuan Zhang, Weiming Hu
Open-vocabulary object detection aims to detect novel object categories beyond the training set.
Ranked #30 on Open Vocabulary Object Detection on MSCOCO
no code implementations • 16 Mar 2022 • Fengxue Zhang, Brian Nord, Yuxin Chen
We show that even with proper network design, such learned representation often leads to collision in the latent space: two points with significantly different observations collide in the learned latent space, leading to degraded optimization performance.
no code implementations • 14 Mar 2022 • Yuling Yan, Gen Li, Yuxin Chen, Jianqing Fan
This paper is concerned with the asynchronous form of Q-learning, which applies a stochastic approximation scheme to Markovian data samples.
no code implementations • 28 Feb 2022 • Laixi Shi, Gen Li, Yuting Wei, Yuxin Chen, Yuejie Chi
Offline or batch reinforcement learning seeks to learn a near-optimal policy using history data without active exploration of the environment.
1 code implementation • 3 Feb 2022 • Himabindu Lakkaraju, Dylan Slack, Yuxin Chen, Chenhao Tan, Sameer Singh
Overall, we hope our work serves as a starting place for researchers and engineers to design interactive explainability systems.
no code implementations • 21 Jan 2022 • Shangrong Yu, Yuxin Chen, Hejun Wu
Low-rank inductive matrix completion (IMC) is currently widely used in IoT data completion, recommendation systems, and so on, as the side information in IMC has demonstrated great potential in reducing sample point remains a major obstacle for the convergence of the nonconvex solutions to IMC.
1 code implementation • 27 Dec 2021 • Renyu Zhang, Aly A. Khan, Robert L. Grossman, Yuxin Chen
To scale up the computation of queries to large batches, we further propose an efficient batch-mode acquisition procedure, which aims to maximize a novel information measure defined through the acquisition function.
no code implementations • NeurIPS 2021 • Akash Kumar, Yuxin Chen, Adish Singla
This learning paradigm has been extensively studied when the learner receives worst-case or random counterexamples; in this paper, we consider the optimal teacher who picks best-case counterexamples to teach the target hypothesis within a hypothesis class.
no code implementations • 3 Nov 2021 • Niklas Åkerblom, Yuxin Chen, Morteza Haghir Chehreghani
In order to learn the model parameters, we develop an online learning framework and investigate several exploration strategies such as Thompson Sampling and Upper Confidence Bound.
no code implementations • 28 Oct 2021 • Matteo Guarrera, Baihong Jin, Tung-Wei Lin, Maria Zuluaga, Yuxin Chen, Alberto Sangiovanni-Vincentelli
We consider the problem of detecting OoD(Out-of-Distribution) input data when using deep neural networks, and we propose a simple yet effective way to improve the robustness of several popular OoD detection methods against label shift.
Out-of-Distribution Detection Out of Distribution (OOD) Detection
no code implementations • NeurIPS 2021 • Chaoqi Wang, Adish Singla, Yuxin Chen
Our focus is to design a teaching algorithm that can provide an informative sequence of contrastive examples to the learner to speed up the learning process.
no code implementations • NeurIPS 2021 • Gen Li, Laixi Shi, Yuxin Chen, Yuejie Chi
Achieving sample efficiency in online episodic reinforcement learning (RL) requires optimally balancing exploration and exploitation.
no code implementations • 12 Aug 2021 • Mengmeng Tian, Yuxin Chen, YuAn Liu, Zehui Xiong, Cyril Leung, Chunyan Miao
It is challenging to design proper incentives for the FL clients due to the fact that the task is privately trained by the clients.
no code implementations • 26 Jul 2021 • Yuling Yan, Yuxin Chen, Jianqing Fan
This paper studies how to construct confidence regions for principal component analysis (PCA) in high dimension, a problem that has been vastly under-explored.
2 code implementations • ICCV 2021 • Yuxin Chen, Ziqi Zhang, Chunfeng Yuan, Bing Li, Ying Deng, Weiming Hu
Graph convolutional networks (GCNs) have been widely used and achieved remarkable results in skeleton-based action recognition.
Ranked #11 on Skeleton Based Action Recognition on N-UCLA
no code implementations • 24 May 2021 • Wenhao Zhan, Shicong Cen, Baihe Huang, Yuxin Chen, Jason D. Lee, Yuejie Chi
These can often be accounted for via regularized RL, which augments the target value function with a structure-promoting regularizer.
no code implementations • NeurIPS 2021 • Gen Li, Yuxin Chen, Yuejie Chi, Yuantao Gu, Yuting Wei
The current paper pertains to a scenario with value-based linear representation, which postulates the linear realizability of the optimal Q-function (also called the "linear $Q^{\star}$ problem").
1 code implementation • 16 May 2021 • Ziyu Ye, Yuxin Chen, Haitao Zheng
We also provide an extensive empirical study on how a biased training anomaly set affects the anomaly score function and therefore the detection performance on different anomaly classes.
no code implementations • 14 Apr 2021 • Chinmaya Mahesh, Kristin Dona, David W. Miller, Yuxin Chen
Data-intensive science is increasingly reliant on real-time processing capabilities and machine learning workflows, in order to filter and analyze the extreme volumes of data being collected.
no code implementations • 7 Apr 2021 • Gen Li, Changxiao Cai, H. Vincent Poor, Yuxin Chen
Eigenvector perturbation analysis plays a vital role in various data science applications.
no code implementations • 22 Feb 2021 • Gen Li, Yuting Wei, Yuejie Chi, Yuxin Chen
The softmax policy gradient (PG) method, which performs gradient ascent under softmax policy parameterization, is arguably one of the de facto implementations of policy optimization in modern reinforcement learning.
no code implementations • 12 Feb 2021 • Gen Li, Changxiao Cai, Yuxin Chen, Yuting Wei, Yuejie Chi
This paper addresses these questions for the synchronous setting: (1) when $|\mathcal{A}|=1$ (so that Q-learning reduces to TD learning), we prove that the sample complexity of TD learning is minimax optimal and scales as $\frac{|\mathcal{S}|}{(1-\gamma)^3\varepsilon^2}$ (up to log factor); (2) when $|\mathcal{A}|\geq 2$, we settle the sample complexity of Q-learning to be on the order of $\frac{|\mathcal{S}||\mathcal{A}|}{(1-\gamma)^4\varepsilon^2}$ (up to log factor).
1 code implementation • 1 Jan 2021 • Ziyu Ye, Yuxin Chen, Haitao Zheng
Given two different anomaly score functions, we formally define their difference in performance as the relative scoring bias of the anomaly detectors.
Semi-supervised Anomaly Detection Supervised Anomaly Detection +1
no code implementations • ICLR 2021 • Ayya Alieva, Aiden Aceves, Jialin Song, Stephen Mayo, Yisong Yue, Yuxin Chen
In particular, we focus on a class of combinatorial problems that can be solved via submodular maximization (either directly on the objective function or via submodular surrogates).
no code implementations • 1 Jan 2021 • Fengxue Zhang, Yair Altas, Louise Fan, Kaustubh Vinchure, Brian Nord, Yuxin Chen
To address this issue, we propose Collision-Free Latent Space Optimization (CoFLO), which employs a novel regularizer to reduce the collision in the learned latent space and encourage the mapping from the latent space to objective value to be Lipschitz continuous.
no code implementations • 15 Dec 2020 • Yuxin Chen, Yuejie Chi, Jianqing Fan, Cong Ma
While the studies of spectral methods can be traced back to classical matrix perturbation theory and methods of moments, the past decade has witnessed tremendous theoretical advances in demystifying their efficacy through the lens of statistical modeling, with the aid of non-asymptotic random matrix theory.
no code implementations • 27 Oct 2020 • Akash Kumar, Hanqi Zhang, Adish Singla, Yuxin Chen
As a warm-up, we show that the teaching complexity is $\Theta(d)$ for the exact teaching of linear perceptrons in $\mathbb{R}^d$, and $\Theta(d^k)$ for kernel perceptron with a polynomial kernel of order $k$.
no code implementations • 17 Oct 2020 • Farnam Mansouri, Yuxin Chen, Ara Vartanian, Xiaojin Zhu, Adish Singla
We analyze several properties of the teaching complexity parameter $TD(\sigma)$ associated with different families of the preference functions, e. g., comparison to the VC dimension of the hypothesis class and additivity/sub-additivity of $TD(\sigma)$ over disjoint domains.
no code implementations • 23 Sep 2020 • Yanxi Chen, Cong Ma, H. Vincent Poor, Yuxin Chen
We study the problem of learning mixtures of low-rank models, i. e. reconstructing multiple low-rank matrices from unlabelled linear measurements of each.
no code implementations • 20 Aug 2020 • Baihong Jin, Yingshui Tan, Albert Liu, Xiangyu Yue, Yuxin Chen, Alberto Sangiovanni Vincentelli
Incipient anomalies present milder symptoms compared to severe ones, and are more difficult to detect and diagnose due to their close resemblance to normal operating conditions.
no code implementations • 4 Aug 2020 • Yuxin Chen, Jianqing Fan, Bingyan Wang, Yuling Yan
We investigate the effectiveness of convex relaxation and nonconvex optimization in solving bilinear systems of equations under two different designs (i. e.$~$a sort of random Fourier design and Gaussian design).
no code implementations • 13 Jul 2020 • Shicong Cen, Chen Cheng, Yuxin Chen, Yuting Wei, Yuejie Chi
This class of methods is often applied in conjunction with entropy regularization -- an algorithmic scheme that encourages exploration -- and is closely related to soft policy iteration and trust region policy optimization.
no code implementations • 12 Jul 2020 • Yingshui Tan, Baihong Jin, Xiangyu Yue, Yuxin Chen, Alberto Sangiovanni Vincentelli
Ensemble learning is widely applied in Machine Learning (ML) to improve model performance and to mitigate decision risks.
no code implementations • 7 Jul 2020 • Baihong Jin, Yingshui Tan, Yuxin Chen, Kameshwar Poolla, Alberto Sangiovanni Vincentelli
Intermediate-Severity (IS) faults present milder symptoms compared to severe faults, and are more difficult to detect and diagnose due to their close resemblance to normal operating conditions.
no code implementations • 25 Jun 2020 • Akash Kumar, Adish Singla, Yisong Yue, Yuxin Chen
We investigate the average teaching complexity of the task, i. e., the minimal number of samples (halfspace queries) required by a teacher to help a version-space learner in locating a randomly selected target.
no code implementations • ICML 2020 • Changxiao Cai, H. Vincent Poor, Yuxin Chen
Furthermore, our findings unveil the statistical optimality of nonconvex tensor completion: it attains un-improvable $\ell_{2}$ accuracy -- including both the rates and the pre-constants -- when estimating both the unknown tensor and the underlying tensor factors.
no code implementations • NeurIPS 2020 • Gen Li, Yuting Wei, Yuejie Chi, Yuantao Gu, Yuxin Chen
Focusing on a $\gamma$-discounted MDP with state space $\mathcal{S}$ and action space $\mathcal{A}$, we demonstrate that the $\ell_{\infty}$-based sample complexity of classical asynchronous Q-learning --- namely, the number of samples needed to yield an entrywise $\varepsilon$-accurate estimate of the Q-function --- is at most on the order of $\frac{1}{\mu_{\min}(1-\gamma)^5\varepsilon^2}+ \frac{t_{mix}}{\mu_{\min}(1-\gamma)}$ up to some logarithmic factor, provided that a proper constant learning rate is adopted.
no code implementations • NeurIPS 2020 • Gen Li, Yuting Wei, Yuejie Chi, Yuxin Chen
This paper is concerned with the sample efficiency of reinforcement learning, assuming access to a generative model (or simulator).
no code implementations • 21 Mar 2020 • Rati Devidze, Farnam Mansouri, Luis Haug, Yuxin Chen, Adish Singla
Machine teaching studies the interaction between a teacher and a student/learner where the teacher selects training examples for the learner to learn a specific task.
no code implementations • 3 Mar 2020 • Niklas Åkerblom, Yuxin Chen, Morteza Haghir Chehreghani
In order to learn the model parameters, we develop an online learning framework and investigate several exploration strategies such as Thompson Sampling and Upper Confidence Bound.
no code implementations • 17 Feb 2020 • Shi Yu, Yuxin Chen, Hussain Zaidi
Our main novel contribution is the discussion about uncertainty measure for BERT, where three different approaches are systematically compared on real problems.
no code implementations • 27 Jan 2020 • Zhe Xu, Yuxin Chen, Ufuk Topcu
In the context of teaching temporal logic formulas, an exhaustive search even for a myopic solution takes exponential time (with respect to the time span of the task).
no code implementations • 15 Jan 2020 • Yuxin Chen, Jianqing Fan, Cong Ma, Yuling Yan
This paper delivers improved theoretical guarantees for the convex programming approach in low-rank matrix estimation, in the presence of (1) random noise, (2) gross sparse outliers, and (3) missing data.
no code implementations • 14 Jan 2020 • Chen Cheng, Yuting Wei, Yuxin Chen
This paper aims to address two fundamental challenges arising in eigenvector estimation and inference for a low-rank matrix from noisy observations: (1) how to estimate an unknown eigenvector when the eigen-gap (i. e. the spacing between the associated eigenvalue and the rest of the spectrum) is particularly small; (2) how to perform estimation and inference on linear functionals of an eigenvector -- a sort of "fine-grained" statistical reasoning that goes far beyond the usual $\ell_2$ analysis.
no code implementations • NeurIPS 2019 • Changxiao Cai, Gen Li, H. Vincent Poor, Yuxin Chen
We study a noisy tensor completion problem of broad practical interest, namely, the reconstruction of a low-rank tensor from highly incomplete and randomly corrupted observations of its entries.
no code implementations • NeurIPS 2019 • Nikhil Ghosh, Yuxin Chen, Yisong Yue
In this paper, we aim to learn a low-dimensional Euclidean representation from a set of constraints of the form "item j is closer to item i than item k".
no code implementations • NeurIPS 2019 • Farnam Mansouri, Yuxin Chen, Ara Vartanian, Xiaojin Zhu, Adish Singla
In our framework, each function $\sigma \in \Sigma$ induces a teacher-learner pair with teaching complexity as $\TD(\sigma)$.
no code implementations • 9 Oct 2019 • Changxiao Cai, Gen Li, Yuejie Chi, H. Vincent Poor, Yuxin Chen
This paper is concerned with estimating the column space of an unknown low-rank matrix $\boldsymbol{A}^{\star}\in\mathbb{R}^{d_{1}\times d_{2}}$, given noisy and partial observations of its entries.
1 code implementation • 12 Sep 2019 • Boyue Li, Shicong Cen, Yuxin Chen, Yuejie Chi
There is growing interest in large-scale machine learning and optimization over decentralized networks, e. g. in the context of multi-agent learning and federated learning.
no code implementations • 10 Sep 2019 • Baihong Jin, Yingshui Tan, Yuxin Chen, Alberto Sangiovanni-Vincentelli
The Monte Carlo dropout method has proved to be a scalable and easy-to-use approach for estimating the uncertainty of deep neural network predictions.
no code implementations • 26 Jul 2019 • Baihong Jin, Yingshui Tan, Alexander Nettekoven, Yuxin Chen, Ufuk Topcu, Yisong Yue, Alberto Sangiovanni Vincentelli
We show that the encoder-decoder model is able to identify the injected anomalies in a modern manufacturing process in an unsupervised fashion.
no code implementations • 10 Jun 2019 • Yuxin Chen, Jianqing Fan, Cong Ma, Yuling Yan
As a byproduct, we obtain a sharp characterization of the estimation accuracy of our de-biased estimators, which, to the best of our knowledge, are the first tractable algorithms that provably achieve full statistical efficiency (including the preconstant).
no code implementations • 17 Apr 2019 • Kevin K. Yang, Yuxin Chen, Alycia Lee, Yisong Yue
Importantly, we show that our objective function can be efficiently decomposed as a difference of submodular functions (DS), which allows us to employ DS optimization tools to greedily identify sets of constraints that increase the likelihood of finding items with high utility.
no code implementations • 28 Mar 2019 • Tian Wang, Zichen Miao, Yuxin Chen, Yi Zhou, Guangcun Shan, Hichem Snoussi
It is challenging to detect the anomaly in crowded scenes for quite a long time.
no code implementations • 20 Feb 2019 • Yuxin Chen, Yuejie Chi, Jianqing Fan, Cong Ma, Yuling Yan
This paper studies noisy low-rank matrix completion: given partial and noisy entries of a large low-rank matrix, the goal is to estimate the underlying matrix faithfully and efficiently.
no code implementations • 18 Feb 2019 • Baihong Jin, Yuxin Chen, Dan Li, Kameshwar Poolla, Alberto Sangiovanni-Vincentelli
The One-Class Support Vector Machine (OC-SVM) is a popular machine learning model for anomaly detection and hence could be used for identifying change points; however, it is sometimes difficult to obtain a good OC-SVM model that can be used on sensor measurement time series to identify the change points in system health status.
no code implementations • 25 Dec 2018 • Yuxin Chen, Morteza Haghir Chehreghani
We propose a novel approach for trip prediction by analyzing user's trip histories.
no code implementations • 30 Nov 2018 • Yuxin Chen, Chen Cheng, Jianqing Fan
The aim is to estimate the leading eigenvalue and eigenvector of $\mathbf{M}^{\star}$.
no code implementations • 15 Nov 2018 • Jialin Song, Yury S. Tokpanov, Yuxin Chen, Dagny Fleischman, Kate T. Fountaine, Harry A. Atwater, Yisong Yue
We apply numerical methods in combination with finite-difference-time-domain (FDTD) simulations to optimize transmission properties of plasmonic mirror color filters using a multi-objective figure of merit over a five-dimensional parameter space by utilizing novel multi-fidelity Gaussian processes approach.
no code implementations • 2 Nov 2018 • Jialin Song, Yuxin Chen, Yisong Yue
How can we efficiently gather information to optimize an unknown function, when presented with multiple, mutually dependent information sources with different costs?
no code implementations • 1 Nov 2018 • Shuangting Liu, Jia-Qi Zhang, Yuxin Chen, Yifan Liu, Zengchang Qin, Tao Wan
Semantic segmentation is one of the basic topics in computer vision, it aims to assign semantic labels to every pixel of an image.
no code implementations • 23 Oct 2018 • Yanzi Zhu, Zhujun Xiao, Yuxin Chen, Zhijing Li, Max Liu, Ben Y. Zhao, Haitao Zheng
Our work demonstrates a new set of silent reconnaissance attacks, which leverages the presence of commodity WiFi devices to track users inside private homes and offices, without compromising any WiFi network, data packets, or devices.
Cryptography and Security
no code implementations • 25 Sep 2018 • Yuejie Chi, Yue M. Lu, Yuxin Chen
Substantial progress has been made recently on developing provably accurate and efficient algorithms for low-rank matrix factorization via nonconvex optimization.
no code implementations • 22 Sep 2018 • Zhujun Xiao, Yanzi Zhu, Yuxin Chen, Ben Y. Zhao, Junchen Jiang, Hai-Tao Zheng
Build accurate DNN models requires training on large labeled, context specific datasets, especially those matching the target scenario.
no code implementations • ICML 2018 • Cong Ma, Kaizheng Wang, Yuejie Chi, Yuxin Chen
Focusing on two statistical estimation problems, i. e. solving random quadratic systems of equations and low-rank matrix completion, we establish that gradient descent achieves near-optimal statistical and computational guarantees without explicit regularization.
no code implementations • NeurIPS 2019 • Anette Hunziker, Yuxin Chen, Oisin Mac Aodha, Manuel Gomez Rodriguez, Andreas Krause, Pietro Perona, Yisong Yue, Adish Singla
Our framework is both generic, allowing the design of teaching schedules for different memory models, and also interactive, allowing the teacher to adapt the schedule to the underlying forgetting mechanisms of the learner.
no code implementations • 21 Mar 2018 • Yuxin Chen, Yuejie Chi, Jianqing Fan, Cong Ma
This paper considers the problem of solving systems of quadratic equations, namely, recovering an object of interest $\mathbf{x}^{\natural}\in\mathbb{R}^{n}$ from $m$ quadratic equations/samples $y_{i}=(\mathbf{a}_{i}^{\top}\mathbf{x}^{\natural})^{2}$, $1\leq i\leq m$.
no code implementations • CVPR 2018 • Oisin Mac Aodha, Shih-An Su, Yuxin Chen, Pietro Perona, Yisong Yue
We study the problem of computer-assisted teaching with explanations.
no code implementations • 17 Feb 2018 • Yuanxin Li, Cong Ma, Yuxin Chen, Yuejie Chi
We consider the problem of recovering low-rank matrices from random rank-one measurements, which spans numerous applications including covariance sketching, phase retrieval, quantum state tomography, and learning shallow polynomial neural networks, among others.
no code implementations • NeurIPS 2018 • Yuxin Chen, Adish Singla, Oisin Mac Aodha, Pietro Perona, Yisong Yue
We highlight that adaptivity does not speed up the teaching process when considering existing models of version space learners, such as "worst-case" (the learner picks the next hypothesis randomly from the version space) and "preference-based" (the learner picks hypothesis according to some global preference).
no code implementations • ICML 2018 • Cong Ma, Kaizheng Wang, Yuejie Chi, Yuxin Chen
Recent years have seen a flurry of activities in designing provably efficient nonconvex procedures for solving statistical estimation problems.
no code implementations • 31 Jul 2017 • Yuxin Chen, Jianqing Fan, Cong Ma, Kaizheng Wang
This paper is concerned with the problem of top-$K$ ranking from pairwise comparisons.
no code implementations • 5 Jun 2017 • Pragya Sur, Yuxin Chen, Emmanuel J. Candès
When used for the purpose of statistical inference, logistic models produce p-values for the regression coefficients by using an approximation to the distribution of the likelihood-ratio test.
no code implementations • 16 Mar 2017 • Yuxin Chen, Jean-Michel Renders, Morteza Haghir Chehreghani, Andreas Krause
We consider the optimal value of information (VoI) problem, where the goal is to sequentially select a set of tests with a minimal cost, so that one can efficiently make the best decision based on the observed outcomes.
no code implementations • 19 Sep 2016 • Yuxin Chen, Emmanuel Candes
We prove that for a broad class of statistical models, the proposed projected power method makes no error---and hence converges to the maximum likelihood estimate---in a suitable regime.
no code implementations • 24 May 2016 • Yuxin Chen, S. Hamed Hassani, Andreas Krause
We consider the Bayesian active learning and experimental design problem, where the goal is to learn the value of some unknown target variable through a sequence of informative, noisy tests.
no code implementations • 11 Feb 2016 • Yuxin Chen, Govinda Kamath, Changho Suh, David Tse
Motivated by applications in domains such as social networks and computational biology, we study the problem of community recovery in graphs with locality.
no code implementations • NeurIPS 2015 • Yuxin Chen, Emmanuel J. Candes
We complement our theoretical study with numerical examples showing that solving random quadratic systems is both computationally and statistically not much harder than solving linear systems of the same size---hence the title of this paper.
no code implementations • 27 Apr 2015 • Yuxin Chen, Changho Suh
To approach this minimax limit, we propose a nearly linear-time ranking scheme, called \emph{Spectral MLE}, that returns the indices of the top-$K$ items in accordance to a careful score estimate.
no code implementations • 6 Apr 2015 • Yuxin Chen, Changho Suh, Andrea J. Goldsmith
In particular, our results isolate a family of \emph{minimum} \emph{channel divergence measures} to characterize the degree of measurement corruption, which together with the size of the minimum cut of $\mathcal{G}$ dictates the feasibility of exact information recovery.
no code implementations • 19 May 2014 • Qixing Huang, Yuxin Chen, Leonidas Guibas
Maximum a posteriori (MAP) inference over discrete Markov random fields is a fundamental task spanning a wide spectrum of real-world applications, which is known to be NP-hard for general graphs.
no code implementations • 24 Feb 2014 • Shervin Javdani, Yuxin Chen, Amin Karbasi, Andreas Krause, J. Andrew Bagnell, Siddhartha Srinivasa
Instead of minimizing uncertainty per se, we consider a set of overlapping decision regions of these hypotheses.
no code implementations • 6 Feb 2014 • Yuxin Chen, Leonidas J. Guibas, Qi-Xing Huang
Joint matching over a collection of objects aims at aggregating information from a large collection of similar instances (e. g. images, graphs, shapes) to improve maps between pairs of them.
no code implementations • 2 Oct 2013 • Yuxin Chen, Yuejie Chi, Andrea Goldsmith
Our method admits universally accurate covariance estimation in the absence of noise, as soon as the number of measurements exceeds the information theoretic limits.
no code implementations • 30 Apr 2013 • Yuxin Chen, Yuejie Chi
The paper explores the problem of \emph{spectral compressed sensing}, which aims to recover a spectrally sparse signal from a small random subset of its $n$ time domain samples.
no code implementations • 16 Apr 2013 • Yuxin Chen, Yuejie Chi
The paper studies the problem of recovering a spectrally sparse object from a small number of time domain samples.