1 code implementation • 6 Oct 2024 • Yijiong Yu, Ma Xiufa, Fang Jianwei, Zhi Xu, Su Guangyao, Wang Jiancheng, Yongfeng Huang, Zhixiao Qi, Wei Wang, Weifeng Liu, Ran Chen, Ji Pei
Long-context language models (LCLMs), characterized by their extensive context window, are becoming increasingly popular.
1 code implementation • 1 Oct 2024 • Zhi Xu, Shaozhe Hao, Kai Han
In this paper, we study a new and challenging task, customized concept decomposition, wherein the objective is to leverage diffusion models to decompose a single image and generate visual concepts from various perspectives.
no code implementations • 8 Mar 2024 • Zhi Xu, Dingkang Yang, Mingcheng Li, Yuzheng Wang, Zhaoyu Chen, Jiawei Chen, Jinjie Wei, Lihua Zhang
Human multimodal language understanding (MLU) is an indispensable component of expression analysis (e. g., sentiment or humor) from heterogeneous modalities, including visual postures, linguistic contents, and acoustic behaviours.
1 code implementation • NeurIPS 2023 • Han Liu, Zhi Xu, Xiaotong Zhang, Feng Zhang, Fenglong Ma, Hongyang Chen, Hong Yu, Xianchao Zhang
Black-box hard-label adversarial attack on text is a practical and challenging task, as the text data space is inherently discrete and non-differentiable, and only the predicted label is accessible.
1 code implementation • ICCV 2023 • Dingkang Yang, Shuai Huang, Zhi Xu, Zhenpeng Li, Shunli Wang, Mingcheng Li, Yuzheng Wang, Yang Liu, Kun Yang, Zhaoyu Chen, Yan Wang, Jing Liu, Peixuan Zhang, Peng Zhai, Lihua Zhang
Driver distraction has become a significant cause of severe traffic accidents over the past decade.
1 code implementation • 19 Jul 2022 • Yunhao Ge, Yao Xiao, Zhi Xu, Xingrui Wang, Laurent Itti
We use human experiments to confirm that both HVE and humans predominantly use some specific features to support the classification of specific classes (e. g., texture is the dominant feature to distinguish a zebra from other quadrupeds, both for humans and HVE).
no code implementations • 6 Dec 2021 • Yunhao Ge, Zhi Xu, Yao Xiao, Gan Xin, Yunkui Pang, Laurent Itti
(2) They lack convexity constraints, which is important for meaningfully manipulating specific attributes for downstream tasks.
no code implementations • 29 Sep 2021 • Yunhao Ge, Yao Xiao, Zhi Xu, Linwei Li, Ziyan Wu, Laurent Itti
Take image classification as an example, HNI visualizes the reasoning logic of a NN with class-specific Structural Concept Graphs (c-SCG), which are human-interpretable.
no code implementations • CVPR 2021 • Yunhao Ge, Yao Xiao, Zhi Xu, Meng Zheng, Srikrishna Karanam, Terrence Chen, Laurent Itti, Ziyan Wu
Despite substantial progress in applying neural networks (NN) to a wide variety of areas, they still largely suffer from a lack of transparency and interpretability.
no code implementations • NeurIPS 2021 • Anish Agarwal, Abdullah Alomar, Varkey Alumootil, Devavrat Shah, Dennis Shen, Zhi Xu, Cindy Yang
We consider offline reinforcement learning (RL) with heterogeneous agents under severe data scarcity, i. e., we only observe a single historical trajectory for every agent under an unknown, potentially sub-optimal policy.
no code implementations • 1 Jan 2021 • Yunhao Ge, Gan Xin, Zhi Xu, Yao Xiao, Yunkui Pang, Yining HE, Laurent Itti
DEAE can become a generative model and synthesis semantic controllable samples by interpolating latent code, which can even synthesis novel attribute value never is shown in the original dataset.
1 code implementation • NeurIPS 2020 • Yuzhe Yang, Zhi Xu
Real-world data often exhibits long-tailed distributions with heavy class imbalance, posing great challenges for deep recognition models.
Ranked #25 on Long-tail Learning on CIFAR-100-LT (ρ=50)
no code implementations • NeurIPS 2020 • Devavrat Shah, Dogyoon Song, Zhi Xu, Yuzhe Yang
As our key contribution, we develop a simple, iterative learning algorithm that finds $\epsilon$-optimal $Q$-function with sample complexity of $\widetilde{O}(\frac{1}{\epsilon^{\max(d_1, d_2)+2}})$ when the optimal $Q$-function has low rank $r$ and the discounting factor $\gamma$ is below a certain threshold.
no code implementations • L4DC 2020 • Devavrat Shah, Qiaomin Xie, Zhi Xu
As a proof of concept, we propose an RL policy using Sparse-Sampling-based Monte Carlo Oracle and argue that it satisfies the stability property as long as the system dynamics under the optimal policy respects a Lyapunov function.
no code implementations • 25 Feb 2020 • Devavrat Shah, Varun Somani, Qiaomin Xie, Zhi Xu
For a concrete instance of EIS where random policy is used for "exploration", Monte-Carlo Tree Search is used for "policy improvement" and Nearest Neighbors is used for "supervised learning", we establish that this method finds an $\varepsilon$-approximate value function of Nash equilibrium in $\widetilde{O}(\varepsilon^{-(d+4)})$ steps when the underlying state-space of the game is continuous and $d$-dimensional.
1 code implementation • ICLR 2020 • Yuzhe Yang, Guo Zhang, Zhi Xu, Dina Katabi
In this paper, we propose to exploit the underlying structures of the state-action value function, i. e., Q function, for both planning and deep RL.
1 code implementation • 28 May 2019 • Yuzhe Yang, Guo Zhang, Dina Katabi, Zhi Xu
We show that this process destroys the adversarial structure of the noise, while re-enforcing the global structure in the original image.
no code implementations • ICLR 2019 • Ravichandra Addanki, Mohammad Alizadeh, Shaileshh Bojja Venkatakrishnan, Devavrat Shah, Qiaomin Xie, Zhi Xu
AlphaGo Zero (AGZ) introduced a new {\em tabula rasa} reinforcement learning algorithm that has achieved superhuman performance in the games of Go, Chess, and Shogi with no prior knowledge other than the rules of the game.
no code implementations • 14 Feb 2019 • Devavrat Shah, Qiaomin Xie, Zhi Xu
In effect, we establish that to learn an $\varepsilon$ approximation of the value function with respect to $\ell_\infty$ norm, MCTS combined with nearest neighbor requires a sample size scaling as $\widetilde{O}\big(\varepsilon^{-(d+4)}\big)$, where $d$ is the dimension of the state space.
no code implementations • 19 Dec 2018 • Idan Amit, John Matherly, William Hewlett, Zhi Xu, Yinnon Meshi, Yigal Weinberger
We present cyber-security problems of high importance.
no code implementations • 6 May 2018 • John N. Tsitsiklis, Kuang Xu, Zhi Xu
We formulate a private learning model to study an intrinsic tradeoff between privacy and query complexity in sequential learning.
no code implementations • 27 Feb 2018 • Zhi Xu, Chengtao Li, Stefanie Jegelka
We explore a notion of robustness for generative adversarial models that is pertinent to their internal interactive structure, and show that, perhaps surprisingly, the GAN in its original form is not robust.