no code implementations • 19 Mar 2025 • Yifan Zhang, Chen Huang, Zachary Karas, Dung Thuy Nguyen, Kevin Leach, Yu Huang
Human attention provides valuable yet underexploited signals for code LLM training, offering a perspective beyond purely machine-driven attention.
no code implementations • 25 Jan 2025 • Jieming Cao, Chen Huang, Yanan Zhang, Ruibo Deng, Jincheng Zhang, Wenqiang Lei
Stigma has emerged as one of the major obstacles to effectively diagnosing depression, as it prevents users from open conversations about their struggles.
no code implementations • 15 Jan 2025 • Junling Li, Cheng-Xiang Wang, Chen Huang, Tianrun Qi, Tong Wu
Different from traditional offline channel modeling, digital twin online channel modeling can sense and accurately characterize dynamic wireless channels in real time, and can therefore greatly assist 6G network optimization.
no code implementations • 10 Jan 2025 • Chen Huang, Yang Deng, Wenqiang Lei, Jiancheng Lv, Tat-Seng Chua, Jimmy Xiangji Huang
With the advancement of large language models (LLMs), intelligent models have evolved from mere tools to autonomous agents with their own goals and strategies for cooperating with humans.
no code implementations • 2 Jan 2025 • Youcheng Huang, Chen Huang, Duanyu Feng, Wenqiang Lei, Jiancheng Lv
Understanding the inner workings of Large Language Models (LLMs) is a critical research frontier.
1 code implementation • 4 Dec 2024 • Yiqin Zhang, Qingkui Chen, Chen Huang, Zhengjie Zhang, Meiling Chen, Zhibing Fu
Most data-driven models for medical image analysis rely on universal augmentations to improve performance.
1 code implementation • 2 Dec 2024 • Qianlong Li, Chen Huang, Shuai Li, Yuanxin Xiang, Deng Xiong, Wenqiang Lei
Despite considerable progress having been made in the LLM era, the reasoning processes of existing methods are often implicit, feeding the entire table into prompts, making it difficult to effectively filter out irrelevant information in the table.
no code implementations • 31 Oct 2024 • Chen Huang, Skyler Seto, Samira Abnar, David Grangier, Navdeep Jaitly, Josh Susskind
Then we jointly train a prompt generator, optimized to produce a prompt embedding that stays close to the aggregated summary while minimizing task loss at the same time.
no code implementations • 14 Oct 2024 • Dejia Xu, Yifan Jiang, Chen Huang, Liangchen Song, Thorsten Gernoth, Liangliang Cao, Zhangyang Wang, Hao Tang
Recent studies have attempted to incorporate camera control into the generation process, but their results are often limited to simple trajectories or lack the ability to generate consistent videos from multiple distinct camera paths for the same scene.
1 code implementation • 30 Sep 2024 • Chen Huang, Guoxiu He
Second, after integrating similar labels generated by the LLM, we prompt the LLM to assign the most appropriate label to each sample in the dataset.
1 code implementation • 22 Sep 2024 • Peixin Qin, Chen Huang, Yang Deng, Wenqiang Lei, Tat-Seng Chua
With the aid of large language models, current conversational recommender system (CRS) has gaining strong abilities to persuade users to accept recommended items.
no code implementations • 5 Sep 2024 • Yong Lin, Skyler Seto, Maartje ter Hoeve, Katherine Metcalf, Barry-John Theobald, Xuan Wang, Yizhe Zhang, Chen Huang, Tong Zhang
These findings highlight that DPORM has limited generalization ability and substantiates the integration of an explicit reward model in iterative DPO approaches.
no code implementations • 3 Jul 2024 • Etai Littwin, Omid Saremi, Madhu Advani, Vimal Thilak, Preetum Nakkiran, Chen Huang, Joshua Susskind
A recent successful approach that falls under the JEPA framework is self-distillation, where an online encoder is trained to predict the output of the target encoder, sometimes using a lightweight predictor network.
1 code implementation • 1 Jul 2024 • Yuxuan Wang, Yijun Liu, Fei Yu, Chen Huang, Kexin Li, Zhiguo Wan, Wanxiang Che
Our in-depth category-level analysis reveals a lack of Chinese cultural knowledge in existing VLMs.
1 code implementation • 12 Jun 2024 • Duanyu Feng, Bowen Qin, Chen Huang, Youcheng Huang, Zheng Zhang, Wenqiang Lei
By leveraging this safety direction, Legend can then leverage the semantic distances of paired responses along this direction to annotate margins automatically.
1 code implementation • 20 May 2024 • Tong Zhang, Peixin Qin, Yang Deng, Chen Huang, Wenqiang Lei, Junhong Liu, dingnan jin, Hongru Liang, Tat-Seng Chua
To this end, we introduce CLAMBER, a benchmark for evaluating LLMs using a well-organized taxonomy.
no code implementations • 20 May 2024 • Chen Huang, Yiping Jin, Ilija Ilievski, Wenqiang Lei, Jiancheng Lv
To address this issue, interactive data annotation utilizes an annotation model to provide suggestions for humans to approve or correct.
no code implementations • 20 May 2024 • Yue Chen, Chen Huang, Yang Deng, Wenqiang Lei, dingnan jin, Jia Liu, Tat-Seng Chua
However, they still struggle to deliver promising performance on unseen domains, struggling to achieve effective domain transferability.
no code implementations • 20 May 2024 • Chen Huang, Yang Deng, Wenqiang Lei, Jiancheng Lv, Ido Dagan
As such, informative or hard data is assigned to the expert for annotation, while easy data is handled by the model.
no code implementations • 16 May 2024 • Chen Huang, Xinwei Yang, Yang Deng, Wenqiang Lei, Jiancheng Lv, Tat-Seng Chua
However, successful legal case matching requires the tacit knowledge of legal practitioners, which is difficult to verbalize and encode into machines.
no code implementations • 6 Apr 2024 • Duanyu Feng, Bowen Qin, Chen Huang, Zheng Zhang, Wenqiang Lei
Direct Preference Optimization (DPO), which derives reward signals directly from pairwise preference data, has shown its effectiveness on aligning Large Language Models (LLMs) with human preferences.
1 code implementation • 4 Apr 2024 • Chen Huang, Peixin Qin, Yang Deng, Wenqiang Lei, Jiancheng Lv, Tat-Seng Chua
The conversational recommendation system (CRS) has been criticized regarding its user experience in real-world scenarios, despite recent significant progress achieved in academia.
no code implementations • 11 Mar 2024 • Tong Zhang, Chen Huang, Yang Deng, Hongru Liang, Jia Liu, Zujie Wen, Wenqiang Lei, Tat-Seng Chua
We investigate non-collaborative dialogue agents, which are expected to engage in strategic conversations with diverse users, for securing a mutual agreement that leans favorably towards the system's objectives.
1 code implementation • 1 Feb 2024 • Victor Chernozhukov, Iván Fernández-Val, Chen Huang, Weining Wang
We show that weak dependence along the panel's time series dimension naturally implies approximate sparsity of the most informative moment conditions, motivating the following approach to remove the bias: First, apply LASSO to the cross-section data at each time period to construct most informative (and cross-fitted) instruments, using lagged values of suitable covariates.
1 code implementation • 29 Jan 2024 • Yuhang Zang, Hanlin Goh, Josh Susskind, Chen Huang
Existing vision-language models exhibit strong generalization on a variety of visual domains and tasks.
1 code implementation • 26 Jan 2024 • Chen Huang, Haoyang Li, Yifan Zhang, Wenqiang Lei, Jiancheng Lv
To this end, various methods have been proposed to create an adaptive filter by incorporating an extra filter (e. g., a high-pass filter) extracted from the graph topology.
1 code implementation • 23 Jan 2024 • Chen Huang, Duanyu Feng, Wenqiang Lei, Jiancheng Lv
Motivated by this, we develop a time-efficient approach called DREditor to edit the matching rule of an off-the-shelf dense retrieval model to suit a specific domain.
1 code implementation • 12 Dec 2023 • Chen Huang, Peixin Qin, Wenqiang Lei, Jiancheng Lv
One of the key factors in language productivity and human cognition is the ability of systematic compositionality, which refers to understanding composed unseen examples of seen primitives.
no code implementations • 7 Dec 2023 • Vimal Thilak, Chen Huang, Omid Saremi, Laurent Dinh, Hanlin Goh, Preetum Nakkiran, Joshua M. Susskind, Etai Littwin
In this paper, we introduce LiDAR (Linear Discriminant Analysis Rank), a metric designed to measure the quality of representations within JE architectures.
no code implementations • 13 Oct 2023 • Samira Abnar, Omid Saremi, Laurent Dinh, Shantel Wilson, Miguel Angel Bautista, Chen Huang, Vimal Thilak, Etai Littwin, Jiatao Gu, Josh Susskind, Samy Bengio
We investigate how the use of a mechanism for adaptive and modular computation in transformers facilitates the learning of tasks that demand generalization over the number of sequential computation steps (i. e., the depth of the computation graph).
no code implementations • 9 Sep 2023 • Chen Huang, Judy S. Kim, Angus I. Kirkland
There is a clear need for developments in characterisation techniques that provide detailed information about structure-function relationships in biology.
1 code implementation • 28 Jun 2023 • Xavier Suau, Federico Danieli, T. Anderson Keller, Arno Blaas, Chen Huang, Jason Ramapuram, Dan Busbridge, Luca Zappella
We propose 2D strUctured and EquivarianT representations (coined DUET), which are 2d representations organized in a matrix structure, and equivariant with respect to transformations acting on the input data.
no code implementations • 24 May 2023 • Yuhang Zang, Kaiyang Zhou, Chen Huang, Chen Change Loy
This paper focuses on long-tailed object detection in the semi-supervised learning setting, which poses realistic challenges, but has rarely been studied in the literature.
no code implementations • 7 Mar 2023 • Chen Huang, Hanlin Goh, Jiatao Gu, Josh Susskind
We do so by Masked Augmentation Subspace Training (or MAST) to encode in the single feature space the priors from different data augmentations in a factorized way.
1 code implementation • 13 Oct 2022 • Yuhang Zang, Wei Li, Kaiyang Zhou, Chen Huang, Chen Change Loy
Prompt tuning, a parameter- and data-efficient transfer learning paradigm that tunes only a small number of parameters in a model's input space, has become a trend in the vision community since the emergence of large vision-language models like CLIP.
no code implementations • 11 Oct 2022 • Yifan Zhang, Chen Huang, Yueke Zhang, Kevin Cao, Scott Thomas Andersen, Huajie Shao, Kevin Leach, Yu Huang
We further analyze the impact of human-written and synthetic comments on binary code comprehension tasks, revealing a significant performance disparity.
no code implementations • 22 Aug 2022 • Yang Wu, Yinghua Wang, Jie Huang, Cheng-Xiang Wang, Chen Huang
Due to the indoor none-line-of-sight (NLoS) propagation and multi-access interference (MAI), it is a great challenge to achieve centimeter-level positioning accuracy in indoor scenarios.
1 code implementation • 15 Jul 2022 • Shuangfei Zhai, Navdeep Jaitly, Jason Ramapuram, Dan Busbridge, Tatiana Likhomanenko, Joseph Yitan Cheng, Walter Talbott, Chen Huang, Hanlin Goh, Joshua Susskind
This pretraining strategy which has been used in BERT models in NLP, Wav2Vec models in Speech and, recently, in MAE models in Vision, forces the model to learn about relationships between the content in different parts of the input using autoencoding related objectives.
no code implementations • 5 Jul 2022 • Chen Huang, Walter Talbott, Navdeep Jaitly, Josh Susskind
Inspired by the success of ConvNets that are combined with pooling to capture long-range dependencies, we learn to pool neighboring features for each token before computing attention in a given attention layer.
no code implementations • 31 Mar 2022 • Huahuan Zheng, Keyu An, Zhijian Ou, Chen Huang, Ke Ding, Guanglu Wan
Based on the DR method, we propose a low-order density ratio method (LODR) by replacing the estimation with a low-order weak language model.
4 code implementations • 22 Mar 2022 • Yuhang Zang, Wei Li, Kaiyang Zhou, Chen Huang, Chen Change Loy
To this end, we propose a novel open-vocabulary detector based on DETR -- hence the name OV-DETR -- which, once trained, can detect any object given its class name or an exemplar image.
Ranked #25 on
Open Vocabulary Object Detection
on MSCOCO
no code implementations • 24 Nov 2021 • Chen Huang, Ruisi He, Bo Ai, Andreas F. Molisch, Buon Kiong Lau, Katsuyuki Haneda, Bo Liu, Cheng-Xiang Wang, Mi Yang, Claude Oestges, Zhangdui Zhong
To provide higher data rates, as well as better coverage, cost efficiency, security, adaptability, and scalability, the 5G and beyond 5G networks are developed with various artificial intelligence techniques.
no code implementations • 24 Nov 2021 • Chen Huang, Ruisi He, Bo Ai, Andreas F. Molisch, Buon Kiong Lau, Katsuyuki Haneda, Bo Liu, Cheng-Xiang Wang, Mi Yang, Claude Oestges, Zhangdui Zhong
This two-part paper investigates the application of artificial intelligence (AI) and in particular machine learning (ML) to the study of wireless propagation channels.
no code implementations • 20 Oct 2021 • Chen Huang, Rui Wang, Cheng-Xiang Wang, Pan Tang, Andreas F. Molisch
We validate this model by contrasting the root-mean-square delay spread and the angular spreads of departure/arrival derived from the channel model with the outcomes directly derived from the measurements.
no code implementations • 29 Sep 2021 • Shuangfei Zhai, Walter Talbott, Nitish Srivastava, Chen Huang, Hanlin Goh, Ruixiang Zhang, Joshua M. Susskind
We introduce Dot Product Attention Free Transformer (DAFT), an efficient variant of Transformers \citep{transformer} that eliminates the query-key dot product in self attention.
Ranked #678 on
Image Classification
on ImageNet
11 code implementations • 28 May 2021 • Shuangfei Zhai, Walter Talbott, Nitish Srivastava, Chen Huang, Hanlin Goh, Ruixiang Zhang, Josh Susskind
We introduce Attention Free Transformer (AFT), an efficient variant of Transformers that eliminates the need for dot product self attention.
no code implementations • 16 May 2021 • Victor Chernozhukov, Chen Huang, Weining Wang
We propose employing a debiased-regularized, high-dimensional generalized method of moments (GMM) framework to perform inference on large-scale spatial panel networks.
no code implementations • CVPR 2021 • Chen Huang, Shuangfei Zhai, Pengsheng Guo, Josh Susskind
This leads to consistent improvements since the value function provides effective metric supervision during finetuning, and helps to correct the potential bias of loss-only supervision.
1 code implementation • ICCV 2021 • Yuhang Zang, Chen Huang, Chen Change Loy
We propose a simple yet effective method, Feature Augmentation and Sampling Adaptation (FASA), that addresses the data scarcity issue by augmenting the feature space especially for rare classes.
no code implementations • ACL 2020 • Jie Tan, Changlin Yang, Ying Li, Siliang Tang, Chen Huang, Yueting Zhuang
Measuring the scholarly impact of a document without citations is an important and challenging problem.
no code implementations • 15 May 2019 • Chen Huang, Shuangfei Zhai, Walter Talbott, Miguel Angel Bautista, Shih-Yu Sun, Carlos Guestrin, Josh Susskind
In most machine learning training paradigms a fixed, often handcrafted, loss function is assumed to be a good proxy for an underlying evaluation metric.
1 code implementation • CVPR 2019 • Yining Li, Chen Huang, Chen Change Loy
Unlike existing methods, we propose to estimate dense and intrinsic 3D appearance flow to better guide the transfer of pixels between poses.
no code implementations • ECCV 2018 • Ceyuan Yang, Zhe Wang, Xinge Zhu, Chen Huang, Jianping Shi, Dahua Lin
Human pose, on the other hand, can represent motion patterns intrinsically and interpretably, and impose the geometric constraints regardless of appearance.
1 code implementation • 1 Jun 2018 • Chen Huang, Yining Li, Chen Change Loy, Xiaoou Tang
Data for face analysis often exhibit highly-skewed class distribution, i. e., most data belong to a few majority classes, while the minority classes only contain a scarce amount of instances.
no code implementations • 24 Apr 2018 • Ruoqi Sun, Chen Huang, Jianping Shi, Lizhuang Ma
The task of face attribute manipulation has found increasing applications, but still remains challeng- ing with the requirement of editing the attributes of a face image while preserving its unique details.
no code implementations • 7 Dec 2017 • Chen Huang, Chen Kong, Simon Lucey
Stochastic Gradient Descent (SGD) is the central workhorse for training modern CNNs.
no code implementations • ICCV 2017 • Yining Li, Chen Huang, Xiaoou Tang, Chen-Change Loy
In particular, each tuple consists of a pair of images and 4. 6 discriminative questions (as positive samples) and 5. 9 non-discriminative questions (as negative samples) on average.
no code implementations • ICCV 2017 • Chen Huang, Simon Lucey, Deva Ramanan
Our fundamental insight is to take an adaptive approach, where easy frames are processed with cheap features (such as pixel values), while challenging frames are processed with invariant but expensive deep features.
1 code implementation • ICCV 2017 • Hamed Kiani Galoogahi, Ashton Fagg, Chen Huang, Deva Ramanan, Simon Lucey
In this paper, we propose the first higher frame rate video dataset (called Need for Speed - NfS) and benchmark for visual object tracking.
no code implementations • NeurIPS 2016 • Chen Huang, Chen Change Loy, Xiaoou Tang
Existing deep embedding methods in vision tasks are capable of learning a compact Euclidean space from images, where Euclidean distances correspond to a similarity metric.
Ranked #27 on
Metric Learning
on CUB-200-2011
2 code implementations • 19 Jul 2016 • Guanghan Ning, Zhi Zhang, Chen Huang, Zhihai He, Xiaobo Ren, Haohong Wang
In this paper, we develop a new approach of spatially supervised recurrent convolutional neural networks for visual object tracking.
no code implementations • CVPR 2016 • Chen Huang, Chen Change Loy, Xiaoou Tang
Attributes offer useful mid-level features to interpret visual data.
no code implementations • CVPR 2016 • Chen Huang, Yining Li, Chen Change Loy, Xiaoou Tang
We further demonstrate that more discriminative deep representation can be learned by enforcing a deep network to maintain both inter-cluster and inter-class margins.
no code implementations • 3 Feb 2016 • Chen Huang, Chen Change Loy, Xiaoou Tang
These methods further deteriorate on small, imbalanced data that has a large degree of class overlap.