no code implementations • ICML 2020 • Yanxi Li, Minjing Dong, Yunhe Wang, Chang Xu
This paper searches for the optimal neural architecture by minimizing a proxy of validation loss.
no code implementations • ECCV 2020 • Xikun Zhang, Chang Xu, DaCheng Tao
Dropout has been widely adopted to regularize graph convolutional networks (GCNs) by randomly zeroing entries of the node feature vectors and obtains promising performance on various tasks.
no code implementations • Findings (EMNLP) 2021 • Chang Xu, Jun Wang, Francisco Guzmán, Benjamin Rubinstein, Trevor Cohn
NLP models are vulnerable to data poisoning attacks.
no code implementations • 5 Jun 2023 • Chengbin Du, Yanxi Li, Zhongwei Qiu, Chang Xu
Recently, text-to-image models have been thriving.
1 code implementation • 25 May 2023 • Zhiwei Hao, Jianyuan Guo, Kai Han, Han Hu, Chang Xu, Yunhe Wang
The tremendous success of large models trained on extensive datasets demonstrates that scale is a key ingredient in achieving superior results.
1 code implementation • 25 May 2023 • Tao Huang, Yuan Zhang, Mingkai Zheng, Shan You, Fei Wang, Chen Qian, Chang Xu
To address this, we propose to denoise student features using a diffusion model trained by teacher features.
1 code implementation • 23 May 2023 • Linwei Tao, Minjing Dong, Chang Xu
While different variants of focal loss have been explored, it is difficult to find a balance between over-confidence and under-confidence.
1 code implementation • 21 Apr 2023 • Mingkai Zheng, Xiu Su, Shan You, Fei Wang, Chen Qian, Chang Xu, Samuel Albanie
We investigate the potential of GPT-4~\cite{gpt4} to perform Neural Architecture Search (NAS) -- the task of designing effective neural architectures.
1 code implementation • CVPR 2023 • Chang Xu, Jian Ding, Jinwang Wang, Wen Yang, Huai Yu, Lei Yu, Gui-Song Xia
Despite the exploration of adaptive label assignment in recent oriented object detectors, the extreme geometry shape and limited feature of oriented tiny objects still induce severe mismatch and imbalance issues.
Ranked #1 on
Oriented Object Detction
on DOTA 2.0
no code implementations • 31 Mar 2023 • Daochang Liu, Qiyue Li, AnhDung Dinh, Tingting Jiang, Mubarak Shah, Chang Xu
Temporal action segmentation is crucial for understanding long-form videos.
Ranked #1 on
Action Segmentation
on Breakfast
no code implementations • CVPR 2023 • Zhongwei Qiu, Yang Qiansheng, Jian Wang, Haocheng Feng, Junyu Han, Errui Ding, Chang Xu, Dongmei Fu, Jingdong Wang
To handle the variances of objects as time proceeds, a novel scheme of progressive decoding is used to update pose and shape queries at each frame.
no code implementations • 28 Feb 2023 • Merat Rezaei, Saad S. Nagi, Chang Xu, Sarah McIntyre, Hakan Olausson, Gregory J. Gerling
Brushed stimuli are perceived as pleasant when stroked lightly on the skin surface of a touch receiver at certain velocities.
no code implementations • 21 Feb 2023 • Chuyang Zhou, Jiajun Huang, Daochang Liu, Chengbin Du, Siqi Ma, Surya Nepal, Chang Xu
More specifically, knowledge distillation on both the spatial and frequency branches has degraded performance than distillation only on the spatial branch.
1 code implementation • 13 Feb 2023 • Linwei Tao, Minjing Dong, Daochang Liu, Changming Sun, Chang Xu
However, early stopping, as a well-known technique to mitigate overfitting, fails to calibrate networks.
no code implementations • 13 Feb 2023 • Jiajun Huang, Xinqi Zhu, Chengbin Du, Siqi Ma, Surya Nepal, Chang Xu
To enhance the performance for such models, we consider the weak compressed and strong compressed data as two views of the original data and they should have similar representation and relationships with other samples.
1 code implementation • 13 Feb 2023 • Yunke Wang, Bo Du, Chang Xu
The trajectories of an initial agent policy could be closer to those non-optimal expert demonstrations, but within the framework of adversarial imitation learning, agent policy will be optimized to cheat the discriminator and produce trajectories that are similar to those optimal expert demonstrations.
no code implementations • 2 Feb 2023 • Zunzhi You, Daochang Liu, Bohyung Han, Chang Xu
Recent advancements in masked image modeling (MIM) have made it a prevailing framework for self-supervised visual representation learning.
1 code implementation • CVPR 2023 • Minjing Dong, Chang Xu
Deep Neural Networks show superior performance in various tasks but are vulnerable to adversarial attacks.
no code implementations • CVPR 2023 • Chen Chen, Daochang Liu, Siqi Ma, Surya Nepal, Chang Xu
However, apart from this standard utility, we identify the "reversed utility" as another crucial aspect, which computes the accuracy on generated data of a classifier trained using real data, dubbed as real2gen accuracy (r2g%).
no code implementations • CVPR 2023 • Yanxi Li, Chang Xu
Although deep neural networks (DNNs) have shown great successes in computer vision tasks, they are vulnerable to perturbations on inputs, and there exists a trade-off between the natural accuracy and robustness to such perturbations, which is mainly caused by the existence of robust non-predictive features and non-robust predictive features.
1 code implementation • 28 Dec 2022 • Guanzhou Ke, Guoqing Chao, Xiaoli Wang, Chenyang Xu, Chang Xu, Yongqi Zhu, Yang Yu
To this end, we utilize a deep fusion network to fuse view-specific representations into the view-common representation, extracting high-level semantics for obtaining robust representation.
1 code implementation • 27 Dec 2022 • Zhongwei Qiu, Huan Yang, Jianlong Fu, Daochang Liu, Chang Xu, Dongmei Fu
Video Super-Resolution (VSR) aims to restore high-resolution (HR) videos from low-resolution (LR) videos.
Ranked #1 on
Video Super-Resolution
on Vid4 - 4x upscaling
no code implementations • 14 Dec 2022 • Xinqi Zhu, Chang Xu, DaCheng Tao
In this paper, we propose a model that automates this process and achieves state-of-the-art semantic discovery performance.
1 code implementation • 13 Dec 2022 • Jianyuan Guo, Kai Han, Han Wu, Yehui Tang, Yunhe Wang, Chang Xu
This paper presents FastMIM, a simple and generic framework for expediting masked image modeling with the following two steps: (i) pre-training vision backbones with low-resolution input images; and (ii) reconstructing Histograms of Oriented Gradients (HOG) feature instead of original RGB values of the input images.
8 code implementations • 23 Nov 2022 • Yehui Tang, Kai Han, Jianyuan Guo, Chang Xu, Chao Xu, Yunhe Wang
The convolutional operation can only capture local information in a window region, which prevents performance from being further improved.
1 code implementation • 26 Oct 2022 • Haoyu Xie, Changqi Wang, Mingkai Zheng, Minjing Dong, Shan You, Chong Fu, Chang Xu
In prevalent pixel-wise contrastive learning solutions, the model maps pixels to deterministic representations and regularizes them in the latent space.
no code implementations • 3 Sep 2022 • Yingtao Luo, Chang Xu, Yang Liu, Weiqing Liu, Shun Zheng, Jiang Bian
In this work, we propose an learning framework that can automatically obtain interpretable PDE models from sequential data.
1 code implementation • 18 Aug 2022 • Chang Xu, Jinwang Wang, Wen Yang, Huai Yu, Lei Yu, Gui-Song Xia
Then, instead of assigning samples with IoU or center sampling strategy, a new Receptive Field Distance (RFD) is proposed to directly measure the similarity between the Gaussian receptive field and ground truth.
1 code implementation • 12 Jul 2022 • Tao Huang, Lang Huang, Shan You, Fei Wang, Chen Qian, Chang Xu
Vision transformers (ViTs) are usually considered to be less light-weight than convolutional neural networks (CNNs) due to the lack of inductive bias.
1 code implementation • 28 Jun 2022 • Chang Xu, Jinwang Wang, Wen Yang, Huai Yu, Lei Yu, Gui-Song Xia
Tiny object detection (TOD) in aerial images is challenging since a tiny object only contains a few pixels.
1 code implementation • 29 May 2022 • Tao Huang, Yuan Zhang, Shan You, Fei Wang, Chen Qian, Jian Cao, Chang Xu
To obtain a group of masks, the receptive tokens are learned via the regular task loss but with teacher fixed, and we also leverage a Dice loss to enrich the diversity of learned masks.
1 code implementation • 21 May 2022 • Tao Huang, Shan You, Fei Wang, Chen Qian, Chang Xu
In this paper, we show that simply preserving the relations between the predictions of teacher and student would suffice, and propose a correlation-based loss to capture the intrinsic inter-class relations from the teacher explicitly.
Ranked #1 on
Knowledge Distillation
on ImageNet
(using extra training data)
1 code implementation • 25 Mar 2022 • Xiu Su, Shan You, Jiyang Xie, Fei Wang, Chen Qian, ChangShui Zhang, Chang Xu
In BCNet, each channel is fairly trained and responsible for the same amount of network widths, thus each network width can be evaluated more accurately.
1 code implementation • CVPR 2022 • Tao Huang, Shan You, Bohan Zhang, Yuxuan Du, Fei Wang, Chen Qian, Chang Xu
Structural re-parameterization (Rep) methods achieve noticeable improvements on simple VGG-style networks.
no code implementations • 16 Mar 2022 • Mingkai Zheng, Shan You, Fei Wang, Chen Qian, ChangShui Zhang, Xiaogang Wang, Chang Xu
Self-supervised Learning (SSL) including the mainstream contrastive learning has achieved great success in learning visual representations without data annotations.
Ranked #47 on
Self-Supervised Image Classification
on ImageNet
Contrastive Learning
Self-Supervised Image Classification
+1
1 code implementation • CVPR 2022 • Mingkai Zheng, Shan You, Lang Huang, Fei Wang, Chen Qian, Chang Xu
Learning with few labeled data has been a longstanding problem in the computer vision and machine learning research community.
no code implementations • 3 Mar 2022 • Yunke Wang, Bo Du, Chang Xu
To satisfy the sequential input of Transformer, the tail of ViT first splits each image into a sequence of visual tokens with a fixed length.
1 code implementation • ICLR 2022 • Tao Huang, Zekang Li, Hua Lu, Yong Shan, Shusheng Yang, Yang Feng, Fei Wang, Shan You, Chang Xu
Evaluation metrics in machine learning are often hardly taken as loss functions, as they could be non-differentiable and non-decomposable, e. g., average precision and F1 score.
3 code implementations • 10 Jan 2022 • Kai Han, Yunhe Wang, Chang Xu, Jianyuan Guo, Chunjing Xu, Enhua Wu, Qi Tian
The proposed C-Ghost module can be taken as a plug-and-play component to upgrade existing convolutional neural networks.
no code implementations • CVPR 2022 • Xueyu Wang, Jiajun Huang, Siqi Ma, Surya Nepal, Chang Xu
We argue that the detectors do not share a similar perspective as human eyes, which might still be spoofed by the disrupted data.
no code implementations • NeurIPS 2021 • Xinghao Chen, Chang Xu, Minjing Dong, Chunjing Xu, Yunhe Wang
Adder neural networks (AdderNets) have shown impressive performance on image classification with only addition operations, which are more energy efficient than traditional convolutional neural networks built with multiplications.
no code implementations • NeurIPS 2021 • Minjing Dong, Yunhe Wang, Xinghao Chen, Chang Xu
Adder neural network (AdderNet) replaces the original convolutions with massive multiplications by cheap additions while achieving comparable performance thus yields a series of energy-efficient neural networks.
no code implementations • NeurIPS 2021 • Minjing Dong, Yunhe Wang, Xinghao Chen, Chang Xu
Adder neural networks (ANNs) are designed for low energy cost which replace expensive multiplications in convolutional neural networks (CNNs) with cheaper additions to yield energy-efficient neural networks and hardware accelerations.
7 code implementations • CVPR 2022 • Yehui Tang, Kai Han, Jianyuan Guo, Chang Xu, Yanxi Li, Chao Xu, Yunhe Wang
To dynamically aggregate tokens, we propose to represent each token as a wave function with two parts, amplitude and phase.
no code implementations • CVPR 2022 • Tao Huang, Shan You, Fei Wang, Chen Qian, ChangShui Zhang, Xiaogang Wang, Chang Xu
In this paper, we leverage an explicit path filter to capture the characteristics of paths and directly filter those weak ones, so that the search can be thus implemented on the shrunk space more greedily and efficiently.
3 code implementations • 1 Nov 2021 • Guanghua Yu, Qinyao Chang, Wenyu Lv, Chang Xu, Cheng Cui, Wei Ji, Qingqing Dang, Kaipeng Deng, Guanzhong Wang, Yuning Du, Baohua Lai, Qiwen Liu, Xiaoguang Hu, dianhai yu, Yanjun Ma
We investigate the applicability of the anchor-free strategy on lightweight object detection models.
Ranked #1 on
Object Detection
on MSCOCO
2 code implementations • 26 Oct 2021 • Jinwang Wang, Chang Xu, Wen Yang, Lei Yu
Our key observation is that Intersection over Union (IoU) based metrics such as IoU itself and its extensions are very sensitive to the location deviation of the tiny objects, and drastically deteriorate the detection performance when used in anchor-based detectors.
Ranked #1 on
Object Detection
on AI-TOD
1 code implementation • ICCV 2021 • Mingkai Zheng, Fei Wang, Shan You, Chen Qian, ChangShui Zhang, Xiaogang Wang, Chang Xu
Specifically, our proposed framework is based on two projection heads, one of which will perform the regular instance discrimination task.
no code implementations • 20 Sep 2021 • Kai Han, Yunhe Wang, Chang Xu, Chunjing Xu, Enhua Wu, DaCheng Tao
A series of secondary filters can be derived from a primary filter with the help of binary masks.
6 code implementations • CVPR 2022 • Jianyuan Guo, Yehui Tang, Kai Han, Xinghao Chen, Han Wu, Chao Xu, Chang Xu, Yunhe Wang
Previous vision MLPs such as MLP-Mixer and ResMLP accept linearly flattened image patches as input, making them inflexible for different input sizes and hard to capture spatial information.
1 code implementation • 18 Aug 2021 • Jiajun Huang, Xueyu Wang, Bo Du, Pei Du, Chang Xu
It includes 10, 000 facial animation videos in ten different actions, which can spoof the recent liveness detectors.
no code implementations • NeurIPS 2021 • Yanxi Li, Zhaohui Yang, Yunhe Wang, Chang Xu
With the tremendous advances in the architecture and scale of convolutional neural networks (CNNs) over the past few decades, they can easily reach or even exceed the performance of humans in certain tasks.
2 code implementations • NeurIPS 2021 • Mingkai Zheng, Shan You, Fei Wang, Chen Qian, ChangShui Zhang, Xiaogang Wang, Chang Xu
Self-supervised Learning (SSL) including the mainstream contrastive learning has achieved great success in learning visual representations without data annotations.
Ranked #65 on
Self-Supervised Image Classification
on ImageNet
Contrastive Learning
Self-Supervised Image Classification
+1
1 code implementation • Findings (ACL) 2021 • Jun Wang, Chang Xu, Francisco Guzman, Ahmed El-Kishky, Benjamin I. P. Rubinstein, Trevor Cohn
Mistranslated numbers have the potential to cause serious effects, such as financial loss or medical misinformation.
10 code implementations • CVPR 2022 • Jianyuan Guo, Kai Han, Han Wu, Yehui Tang, Xinghao Chen, Yunhe Wang, Chang Xu
Vision transformers have been successfully applied to image recognition tasks due to their ability to capture long-range dependencies within an image.
1 code implementation • 12 Jul 2021 • Jun Wang, Chang Xu, Francisco Guzman, Ahmed El-Kishky, Yuqing Tang, Benjamin I. P. Rubinstein, Trevor Cohn
Neural machine translation systems are known to be vulnerable to adversarial test inputs, however, as we show in this paper, these systems are also vulnerable to training attacks.
4 code implementations • NeurIPS 2021 • Yehui Tang, Kai Han, Chang Xu, An Xiao, Yiping Deng, Chao Xu, Yunhe Wang
Transformer models have achieved great progress on computer vision tasks recently.
1 code implementation • 25 Jun 2021 • Xiu Su, Shan You, Jiyang Xie, Mingkai Zheng, Fei Wang, Chen Qian, ChangShui Zhang, Xiaogang Wang, Chang Xu
Vision transformers (ViTs) inherited the success of NLP but their structures have not been sufficiently investigated and optimized for visual tasks.
1 code implementation • CVPR 2021 • Hanting Chen, Tianyu Guo, Chang Xu, Wenshuo Li, Chunjing Xu, Chao Xu, Yunhe Wang
Experiments on various datasets demonstrate that the student networks learned by the proposed method can achieve comparable performance with those using the original dataset.
no code implementations • CVPR 2021 • Jianyuan Guo, Kai Han, Han Wu, Chao Zhang, Xinghao Chen, Chunjing Xu, Chang Xu, Yunhe Wang
In this paper, we present a positive-unlabeled learning based scheme to expand training data by purifying valuable images from massive unlabeled ones, where the original training data are viewed as positive data and the unlabeled images in the wild are unlabeled data.
2 code implementations • CVPR 2021 • Yixing Xu, Yunhe Wang, Kai Han, Yehui Tang, Shangling Jui, Chunjing Xu, Chang Xu
An effective and efficient architecture performance evaluation scheme is essential for the success of Neural Architecture Search (NAS).
no code implementations • 11 Jun 2021 • Xiu Su, Shan You, Mingkai Zheng, Fei Wang, Chen Qian, ChangShui Zhang, Chang Xu
The operation weight for each path is represented as a convex combination of items in a dictionary with a simplex code.
1 code implementation • 7 Jun 2021 • Xinqi Zhu, Chang Xu, DaCheng Tao
Instead, we propose to encode the data variations with groups, a structure not only can equivariantly represent variations, but can also be adaptively optimized to preserve the properties of data variations.
no code implementations • CVPR 2022 • Yehui Tang, Kai Han, Yunhe Wang, Chang Xu, Jianyuan Guo, Chao Xu, DaCheng Tao
We first identify the effective patches in the last layer and then use them to guide the patch selection process of previous layers.
no code implementations • 29 May 2021 • Hanting Chen, Yunhe Wang, Chang Xu, Chao Xu, Chunjing Xu, Tong Zhang
The widely-used convolutions in deep neural networks are exactly cross-correlation to measure the similarity between input feature and convolution filters, which involves massive multiplications between float values.
no code implementations • CVPR 2021 • Xiu Su, Shan You, Fei Wang, Chen Qian, ChangShui Zhang, Chang Xu
In BCNet, each channel is fairly trained and responsible for the same amount of network widths, thus each network width can be evaluated more accurately.
1 code implementation • CVPR 2021 • Xinqi Zhu, Chang Xu, DaCheng Tao
We thus impose a perturbation on a certain dimension of the latent code, and expect to identify the perturbation along this dimension from the generated images so that the encoding of simple variations can be enforced.
1 code implementation • CVPR 2021 • Jianyuan Guo, Kai Han, Yunhe Wang, Han Wu, Xinghao Chen, Chunjing Xu, Chang Xu
To this end, we present a novel distillation algorithm via decoupled features (DeFeat) for learning a better student detector.
no code implementations • 23 Mar 2021 • Jingwei Xu, Siyuan Zhu, Zenan Li, Chang Xu
Specifically, We construct a generative model, called Latent Sequential Gaussian Mixture (LSGM), to depict how the in-distribution latent features are generated in terms of the trace of DNN inference across representation spaces.
Out-of-Distribution Detection
Out of Distribution (OOD) Detection
1 code implementation • CVPR 2021 • Xiu Su, Tao Huang, Yanxi Li, Shan You, Fei Wang, Chen Qian, ChangShui Zhang, Chang Xu
One-shot neural architecture search (NAS) methods significantly reduce the search cost by considering the whole search space as one network, which only needs to be trained once.
no code implementations • ICCV 2021 • Wenbin Xie, Dehua Song, Chang Xu, Chunjing Xu, HUI ZHANG, Yunhe Wang
Extensive experiments conducted on benchmark SISR models and datasets show that the frequency-aware dynamic network can be employed for various SISR neural architectures to obtain the better tradeoff between visual quality and computational complexity.
3 code implementations • CVPR 2021 • Yehui Tang, Yunhe Wang, Yixing Xu, Yiping Deng, Chao Xu, DaCheng Tao, Chang Xu
Then, the manifold relationship between instances and the pruned sub-networks will be aligned in the training procedure.
1 code implementation • NeurIPS 2021 • Yixing Xu, Kai Han, Chang Xu, Yehui Tang, Chunjing Xu, Yunhe Wang
Binary neural networks (BNNs) represent original full-precision weights and activations into 1-bit with sign function.
no code implementations • 1 Mar 2021 • Ziqing Lu, Chang Xu, Bo Du, Takashi Ishida, Lefei Zhang, Masashi Sugiyama
In neural networks, developing regularization algorithms to settle overfitting is one of the major study areas.
no code implementations • 24 Feb 2021 • Ying Wang, Liang Qiao, Chang Xu, Yepang Liu, Shing-Chi Cheung, Na Meng, Hai Yu, Zhiliang Zhu
The results showed that \textsc{Hero} achieved a high detection rate of 98. 5\% on a DM issue benchmark and found 2, 422 new DM issues in 2, 356 popular Golang projects.
Software Engineering
no code implementations • 15 Feb 2021 • Wentao Xu, Weiqing Liu, Chang Xu, Jiang Bian, Jian Yin, Tie-Yan Liu
To remedy the first shortcoming, we propose to model the stock context and learn the effect of event information on the stocks under different contexts.
no code implementations • ICLR 2021 • Xiu Su, Shan You, Tao Huang, Fei Wang, Chen Qian, ChangShui Zhang, Chang Xu
In this paper, to better evaluate each width, we propose a locally free weight sharing strategy (CafeNet) accordingly.
no code implementations • 20 Dec 2020 • Huaxi Huang, Junjie Zhang, Jian Zhang, Qiang Wu, Chang Xu
Second, the extra unlabeled samples are employed to transfer the knowledge from base classes to novel classes through contrastive learning.
no code implementations • 3 Dec 2020 • Niu Wan, Takayuki Myo, Chang Xu, Hiroshi Toki, Hisashi Horiuchi, Mengjiao Lyu
The central short-range correlation coming from the short-range repulsion in the NN interaction is treated by the unitary correlation operator method (UCOM) and the tensor correlation and spin-orbit effects are described by the two-particle two-hole (2p2h) excitations of nucleon pairs, in which the two nucleons with a large relative momentum are regarded as a high-momentum pair (HM).
Nuclear Theory
5 code implementations • CVPR 2021 • Hanting Chen, Yunhe Wang, Tianyu Guo, Chang Xu, Yiping Deng, Zhenhua Liu, Siwei Ma, Chunjing Xu, Chao Xu, Wen Gao
To maximally excavate the capability of transformer, we present to utilize the well-known ImageNet benchmark for generating a large amount of corrupted image pairs.
Ranked #1 on
Single Image Deraining
on Rain100L
(using extra training data)
no code implementations • NeurIPS 2020 • Chu Zhou, Hang Zhao, Jin Han, Chang Xu, Chao Xu, Tiejun Huang, Boxin Shi
A conventional camera often suffers from over- or under-exposure when recording a real-world scene with a very high dynamic range (HDR).
1 code implementation • NeurIPS 2020 • Yanxi Li, Zhaohui Yang, Yunhe Wang, Chang Xu
The power of deep neural networks is to be unleashed for analyzing a large volume of data (e. g. ImageNet), but the architecture search is often executed on another smaller dataset (e. g. CIFAR-10) to finish it in a feasible time.
no code implementations • COLING 2020 • Chang Xu, Cecile Paris, Ross Sparks, Surya Nepal, Keith VanderLinden
Our experimental results show that SIRTA is highly effective in distilling stances from social posts for SLO level assessment, and that the continuous monitoring of SLO levels afforded by SIRTA enables the early detection of critical SLO changes.
no code implementations • 2 Nov 2020 • Chang Xu, Jun Wang, Yuqing Tang, Francisco Guzman, Benjamin I. P. Rubinstein, Trevor Cohn
In this paper, we show that targeted attacks on black-box NMT systems are feasible, based on poisoning a small fraction of their parallel training data.
no code implementations • 28 Oct 2020 • Xiu Su, Shan You, Tao Huang, Hongyan Xu, Fei Wang, Chen Qian, ChangShui Zhang, Chang Xu
To deploy a well-trained CNN model on low-end computation edge devices, it is usually supposed to compress or prune the model under certain computation budget (e. g., FLOPs).
4 code implementations • NeurIPS 2020 • Yehui Tang, Yunhe Wang, Yixing Xu, DaCheng Tao, Chunjing Xu, Chao Xu, Chang Xu
To increase the reliability of the results, we prefer to have a more rigorous research design by including a scientific control group as an essential part to minimize the effect of all factors except the association between the filter and expected network output.
1 code implementation • ICML 2020 • Kai Han, Yunhe Wang, Yixing Xu, Chunjing Xu, Enhua Wu, Chang Xu
This paper formalizes the binarization operations over neural networks from a learning perspective.
no code implementations • 1 Oct 2020 • Zeyu Feng, Chang Xu, DaCheng Tao
Unsupervised open-set domain adaptation (UODA) is a realistic problem where unlabeled target data contain unknown classes.
no code implementations • NeurIPS 2020 • Yixing Xu, Chang Xu, Xinghao Chen, Wei zhang, Chunjing Xu, Yunhe Wang
A convolutional neural network (CNN) with the same architecture is simultaneously initialized and trained as a teacher network, features and weights of ANN and CNN will be transformed to a new space to eliminate the accuracy drop.
1 code implementation • NeurIPS 2020 • Zhaohui Yang, Yunhe Wang, Kai Han, Chunjing Xu, Chao Xu, DaCheng Tao, Chang Xu
Quantized neural networks with low-bit weights and activations are attractive for developing AI accelerators.
no code implementations • CVPR 2021 • Dehua Song, Yunhe Wang, Hanting Chen, Chang Xu, Chunjing Xu, DaCheng Tao
To this end, we thoroughly analyze the relationship between an adder operation and the identity mapping and insert shortcuts to enhance the performance of SR models using adder networks.
no code implementations • 2 Sep 2020 • Minjing Dong, Yanxi Li, Yunhe Wang, Chang Xu
We explore the relationship among adversarial robustness, Lipschitz constant, and architecture parameters and show that an appropriate constraint on architecture parameters could reduce the Lipschitz constant to further improve the robustness.
1 code implementation • ICCV 2019 • Xinqi Zhu, Chang Xu, Langwen Hui, Cewu Lu, DaCheng Tao
Specifically, we show how two-layer subnets in CNNs can be converted to temporal bilinear modules by adding an auxiliary-branch.
1 code implementation • ECCV 2020 • Xinqi Zhu, Chang Xu, DaCheng Tao
Given image pairs generated by latent codes varying in a single dimension, this varied dimension could be closely correlated with these image pairs if the representation is well disentangled.
no code implementations • ECCV 2020 • Xinghao Chen, Yiman Zhang, Yunhe Wang, Han Shu, Chunjing Xu, Chang Xu
This paper proposes to learn a lightweight video style transfer network via knowledge distillation paradigm.
no code implementations • 24 Jun 2020 • Yao Cheng, Chang Xu, Zhen Hai, Yingjiu Li
Moreover, the user study further validates that the generated mnemonic sentences by DeepMnemonic are useful in helping users memorize strong passwords.
2 code implementations • CVPR 2021 • Zhaohui Yang, Yunhe Wang, Xinghao Chen, Jianyuan Guo, Wei zhang, Chao Xu, Chunjing Xu, DaCheng Tao, Chang Xu
To achieve an extremely fast NAS while preserving the high accuracy, we propose to identify the vital blocks and make them the priority in the architecture search.
no code implementations • 28 May 2020 • Huaxi Huang, Jun-Jie Zhang, Jian Zhang, Qiang Wu, Chang Xu
The challenges of high intra-class variance yet low inter-class fluctuations in fine-grained visual categorization are more severe with few labeled samples, \textit{i. e.,} Fine-Grained categorization problems under the Few-Shot setting (FGFS).
no code implementations • CVPR 2020 • Yehui Tang, Yunhe Wang, Yixing Xu, Hanting Chen, Chunjing Xu, Boxin Shi, Chao Xu, Qi Tian, Chang Xu
A graph convolutional neural network is introduced to predict the performance of architectures based on the learned representations and their relation modeled by the graph.
no code implementations • 24 Apr 2020 • Tao Wang, Junsong Wang, Chang Xu, Chao Xue
With the best searched quantization policy, we subsequently retrain or finetune to further improve the performance of the quantized target network.
1 code implementation • CVPR 2020 • Jianyuan Guo, Kai Han, Yunhe Wang, Chao Zhang, Zhaohui Yang, Han Wu, Xinghao Chen, Chang Xu
To this end, we propose a hierarchical trinity search framework to simultaneously discover efficient architectures for all components (i. e. backbone, neck, and head) of object detector in an end-to-end manner.
1 code implementation • 22 Mar 2020 • Jingxin Liu, Chang Xu, Chang Yin, Weiqiang Wu, You Song
Graph representation learning is a fundamental task in various applications that strives to learn low-dimensional embeddings for nodes that can preserve graph topology information.
no code implementations • 13 Mar 2020 • Chang Xu, Cecile Paris, Surya Nepal, Ross Sparks, Chong Long, Yafang Wang
We address the issue of having a limited number of annotations for stance classification in a new domain, by adapting out-of-domain classifiers with domain adaptation.
no code implementations • 7 Mar 2020 • Hanting Chen, Yunhe Wang, Han Shu, Changyuan Wen, Chunjing Xu, Boxin Shi, Chao Xu, Chang Xu
To promote the capability of student generator, we include a student discriminator to measure the distances between real images, and images generated by student and teacher generators.
2 code implementations • 23 Feb 2020 • Yehui Tang, Yunhe Wang, Yixing Xu, Boxin Shi, Chao Xu, Chunjing Xu, Chang Xu
On one hand, massive trainable parameters significantly enhance the performance of these deep networks.
no code implementations • 17 Feb 2020 • Zhaohui Yang, Yunhe Wang, Chang Xu, Peng Du, Chao Xu, Chunjing Xu, Qi Tian
Experiments on benchmarks demonstrate that images compressed by using the proposed method can also be well recognized by subsequent visual recognition and detection models.
1 code implementation • CVPR 2020 • Tianyu Guo, Chang Xu, Jiajun Huang, Yunhe Wang, Boxin Shi, Chao Xu, DaCheng Tao
In contrast, it is more reasonable to treat the generated data as unlabeled, which could be positive or negative according to their quality.
2 code implementations • CVPR 2020 • Hanting Chen, Yunhe Wang, Chunjing Xu, Boxin Shi, Chao Xu, Qi Tian, Chang Xu
The widely-used convolutions in deep neural networks are exactly cross-correlation to measure the similarity between input feature and convolution filters, which involves massive multiplications between float values.
no code implementations • NeurIPS 2019 • Tianyu Guo, Chang Xu, Boxin Shi, Chao Xu, DaCheng Tao
A worst-case formulation can be developed over this distribution set, and then be interpreted as a generation task in an adversarial manner.
25 code implementations • CVPR 2020 • Kai Han, Yunhe Wang, Qi Tian, Jianyuan Guo, Chunjing Xu, Chang Xu
Deploying convolutional neural networks (CNNs) on embedded devices is difficult due to the limited memory and computation resources.
Ranked #779 on
Image Classification
on ImageNet
no code implementations • 6 Oct 2019 • Zenan Li, Xiaoxing Ma, Chang Xu, Jingwei Xu, Chun Cao, Jian Lü
Trained DNN models are increasingly adopted as integral parts of software systems, but they often perform deficiently in the field.
3 code implementations • 30 Sep 2019 • Yixing Xu, Yunhe Wang, Kai Han, Yehui Tang, Shangling Jui, Chunjing Xu, Chang Xu
An effective and efficient architecture performance evaluation scheme is essential for the success of Neural Architecture Search (NAS).
1 code implementation • 25 Sep 2019 • Dehua Song, Chang Xu, Xu Jia, Yiyi Chen, Chunjing Xu, Yunhe Wang
Focusing on this issue, we propose an efficient residual dense block search algorithm with multiple objectives to hunt for fast, lightweight and accurate networks for image super-resolution.
2 code implementations • NeurIPS 2019 • Yixing Xu, Yunhe Wang, Hanting Chen, Kai Han, Chunjing Xu, DaCheng Tao, Chang Xu
In practice, only a small portion of the original training set is required as positive examples and more useful training examples can be obtained from the massive unlabeled data on the cloud through a PU classifier with an attention based multi-scale feature extractor.
1 code implementation • CVPR 2020 • Zhaohui Yang, Yunhe Wang, Xinghao Chen, Boxin Shi, Chao Xu, Chunjing Xu, Qi Tian, Chang Xu
Architectures in the population that share parameters within one SuperNet in the latest generation will be tuned over the training dataset with a few epochs.
1 code implementation • 6 Aug 2019 • Kai Han, Yunhe Wang, Yixing Xu, Chunjing Xu, DaCheng Tao, Chang Xu
Existing works used to decrease the number or size of requested convolution filters for a minimum viable CNN on edge devices.
no code implementations • 27 Jul 2019 • Kai Han, Yunhe Wang, Han Shu, Chuanjian Liu, Chunjing Xu, Chang Xu
This paper expands the strength of deep convolutional neural networks (CNNs) to the pedestrian attribute recognition problem by devising a novel attribute aware pooling algorithm.
no code implementations • 27 Jul 2019 • Chuanjian Liu, Yunhe Wang, Kai Han, Chunjing Xu, Chang Xu
Exploring deep convolutional neural networks of high efficiency and low memory usage is very essential for a wide variety of machine learning tasks.
2 code implementations • ICCV 2019 • Han Shu, Yunhe Wang, Xu Jia, Kai Han, Hanting Chen, Chunjing Xu, Qi Tian, Chang Xu
Generative adversarial networks (GANs) have been successfully used for considerable computer vision tasks, especially the image-to-image translation.
no code implementations • 23 Jul 2019 • Dalu Guo, Chang Xu, DaCheng Tao
The question-graph exchanges information between these output nodes from image-graph to amplify the implicit yet important relationship between objects.
Ranked #16 on
Visual Question Answering (VQA)
on VQA v2 test-std
no code implementations • 13 Jul 2019 • Yehui Tang, Shan You, Chang Xu, Boxin Shi, Chao Xu
Specifically, we exploit the unlabeled data to mimic the classification characteristics of giant networks, so that the original capacity can be preserved nicely.
1 code implementation • 6 Jun 2019 • Zenan Li, Xiaoxing Ma, Chang Xu, Chun Cao, Jingwei Xu, Jian Lü
With the increasing adoption of Deep Neural Network (DNN) models as integral parts of software systems, efficient operational testing of DNNs is much in demand to ensure these models' actual performance in field conditions.
no code implementations • ACL 2019 • Chang Xu, Cecile Paris, Surya Nepal, Ross Sparks
We identify agreement and disagreement between utterances that express stances towards a topic of discussion.
no code implementations • 8 Apr 2019 • Yong Luo, DaCheng Tao, Chang Xu, Chao Xu, Hong Liu, Yonggang Wen
In computer vision, image datasets used for classification are naturally associated with multiple labels and comprised of multiple views, because each image may contain several objects (e. g. pedestrian, bicycle and tree) and is properly characterized by multiple visual features (e. g. color, texture and shape).
2 code implementations • 4 Apr 2019 • Xinyuan Chen, Chang Xu, Xiaokang Yang, Li Song, DaCheng Tao
We propose adversarial gated networks (Gated GAN) to transfer multiple styles in a single model.
no code implementations • 4 Apr 2019 • Meng Liu, Chang Xu, Yong Luo, Chao Xu, Yonggang Wen, DaCheng Tao
Feature selection is beneficial for improving the performance of general machine learning tasks by extracting an informative subset from the high-dimensional features.
no code implementations • 4 Apr 2019 • Chang Xu, DaCheng Tao, Chao Xu
In this paper, we propose the Multi-view Intact Space Learning (MISL) algorithm, which integrates the encoded complementary information in multiple views to discover a latent intact representation of the data.
3 code implementations • ICCV 2019 • Hanting Chen, Yunhe Wang, Chang Xu, Zhaohui Yang, Chuanjian Liu, Boxin Shi, Chunjing Xu, Chao Xu, Qi Tian
Learning portable neural networks is very essential for computer vision for the purpose that pre-trained heavy deep models can be well applied on edge devices such as mobile phones and micro sensors.
no code implementations • CVPR 2019 • Dalu Guo, Chang Xu, DaCheng Tao
Afterward, in the second stage, answers with high probability of being correct are re-ranked by synergizing with image and question.
Ranked #57 on
Visual Dialog
on Visual Dialog v1.0 test-std
3 code implementations • 18 Feb 2019 • Yansong Gao, Chang Xu, Derui Wang, Shiping Chen, Damith C. Ranasinghe, Surya Nepal
Since the trojan trigger is a secret guarded and exploited by the attacker, detecting such trojan inputs is a challenge, especially at run-time when models are in active operation.
Cryptography and Security
no code implementations • 17 Dec 2018 • Hanting Chen, Yunhe Wang, Chang Xu, Chao Xu, DaCheng Tao
Experiments on benchmark datasets and well-trained networks suggest that the proposed algorithm is superior to state-of-the-art teacher-student learning methods in terms of computational and storage complexity.
no code implementations • NeurIPS 2018 • Yunhe Wang, Chang Xu, Chunjing Xu, Chao Xu, DaCheng Tao
A series of secondary filters can be derived from a primary filter.
no code implementations • 13 Nov 2018 • Chang Xu, Weiran Huang, Hongwei Wang, Gang Wang, Tie-Yan Liu
In this paper, we propose an improved variant of RNN, Multi-Channel RNN (MC-RNN), to dynamically capture and leverage local semantic structure information.
no code implementations • 30 Jul 2018 • Tianyu Guo, Chang Xu, Shiyi He, Boxin Shi, Chao Xu, DaCheng Tao
In this way, a portable student network with significantly fewer parameters can achieve a considerable accuracy which is comparable to that of teacher network.
1 code implementation • ACL 2018 • Chang Xu, Cecile Paris, Surya Nepal, Ross Sparks
In stance classification, the target on which the stance is made defines the boundary of the task, and a classifier is usually trained for prediction on the same target.
no code implementations • 16 May 2018 • Xikun Zhang, Chang Xu, Xinmei Tian, DaCheng Tao
Considering the complementarity between graph node convolution and graph edge convolution, we additionally construct two hybrid neural networks to combine graph node convolutional neural network and graph edge convolutional neural network using shared intermediate layers.
no code implementations • ECCV 2018 • Xinyuan Chen, Chang Xu, Xiaokang Yang, DaCheng Tao
This paper studies the object transfiguration problem in wild images.
3 code implementations • 1 Mar 2018 • Chaoyue Wang, Chang Xu, Xin Yao, DaCheng Tao
In this paper, we propose a novel GAN framework called evolutionary generative adversarial networks (E-GAN) for stable GAN training and improved generative performance.
1 code implementation • ICML 2017 • Yunhe Wang, Chang Xu, Chao Xu, DaCheng Tao
The filter is then re-configured to establish the mapping from original input to the new compact feature map, and the resulting network can preserve intrinsic information of the original network with significantly fewer parameters, which not only decreases the online memory for launching CNN but also accelerates the computation speed.
no code implementations • 25 Jul 2017 • Yunhe Wang, Chang Xu, Jiayan Qiu, Chao Xu, DaCheng Tao
In contrast to directly recognizing subtle weights or filters as redundant in a given CNN, this paper presents an evolutionary method to automatically eliminate redundant convolution filters.
2 code implementations • 28 Jun 2017 • Chaoyue Wang, Chang Xu, Chaohui Wang, DaCheng Tao
The proposed PAN consists of two feed-forward convolutional neural networks (CNNs), the image transformation network T and the discriminative network D. Through combining the generative adversarial loss and the proposed perceptual adversarial loss, these two networks can be trained alternately to solve image-to-image transformation tasks.
no code implementations • 31 May 2017 • Chang Xu, Tao Qin, Gang Wang, Tie-Yan Liu
Stochastic gradient descent (SGD), which updates the model parameters by adding a local gradient times a learning rate at each step, is widely used in model training of machine learning algorithms such as neural networks.
no code implementations • International Joint Conference on Artificial Intelligence 2017 • Chaoyue Wang, Chaohui Wang, Chang Xu, DaCheng Tao
The whole framework consists of a disentangling network, a generative network, a tag mapping net, and a discriminative network, which are trained jointly based on a given set of images that are complete/partially tagged(i. e., supervised/semi-supervised setting).
no code implementations • 25 Jan 2017 • Shan You, Chang Xu, Yunhe Wang, Chao Xu, DaCheng Tao
This paper presents privileged multi-label learning (PrML) to explore and exploit the relationship between labels in multi-label learning problems.
no code implementations • NeurIPS 2016 • Yunhe Wang, Chang Xu, Shan You, DaCheng Tao, Chao Xu
Deep convolutional neural networks (CNNs) are successfully used in a number of applications.
no code implementations • 28 Apr 2016 • Chang Xu, DaCheng Tao, Chao Xu
An underlying assumption in conventional multi-view learning algorithms is that all views can be simultaneously accessed.
no code implementations • 19 Apr 2016 • Yunhe Wang, Chang Xu, Shan You, DaCheng Tao, Chao Xu
Here we study the extreme visual recovery problem, in which over 90\% of pixel values in a given image are missing.
no code implementations • 19 Apr 2016 • Shan You, Chang Xu, Yunhe Wang, Chao Xu, DaCheng Tao
The core of SLL is to explore and exploit the relationships between new labels and past labels and then inherit the relationship into hypotheses of labels to boost the performance of new classifiers.
no code implementations • 26 Oct 2014 • Chang Xu, Tongliang Liu, DaCheng Tao, Chao Xu
We analyze the local Rademacher complexity of empirical risk minimization (ERM)-based multi-label learning algorithms, and in doing so propose a new algorithm for multi-label learning.
no code implementations • 23 Apr 2014 • Yuyu Zhang, Hanjun Dai, Chang Xu, Jun Feng, Taifeng Wang, Jiang Bian, Bin Wang, Tie-Yan Liu
Click prediction is one of the fundamental problems in sponsored search.
no code implementations • 20 Apr 2013 • Chang Xu, DaCheng Tao, Chao Xu
Notably, co-training style algorithms train alternately to maximize the mutual agreement on two distinct views of the data; multiple kernel learning algorithms exploit kernels that naturally correspond to different views and combine kernels either linearly or non-linearly to improve learning performance; and subspace learning algorithms aim to obtain a latent subspace shared by multiple views by assuming that the input views are generated from this latent subspace.