1 code implementation • 16 Apr 2023 • Wenke Xia, Xingjian Li, Andong Deng, Haoyi Xiong, Dejing Dou, Di Hu
However, such semantic consistency from the synchronization is hard to guarantee in unconstrained videos, due to the irrelevant modality noise and differentiated semantic correlation.
1 code implementation • 14 Jul 2022 • Ji Liu, daxiang dong, Xi Wang, An Qin, Xingjian Li, Patrick Valduriez, Dejing Dou, dianhai yu
Although more layers and more parameters generally improve the accuracy of the models, such big models generally have high computational complexity and require big memory, which exceed the capacity of small devices for inference and incurs long training time.
no code implementations • 12 Jun 2022 • Hang Hua, Xingjian Li, Dejing Dou, Cheng-Zhong Xu, Jiebo Luo
The advent of large-scale pre-trained language models has contributed greatly to the recent progress in natural language processing.
no code implementations • 26 May 2022 • Xingjian Li, Pengkun Yang, Tianyang Wang, Xueying Zhan, Min Xu, Dejing Dou, Chengzhong Xu
Uncertainty estimation for unlabeled data is crucial to active learning.
no code implementations • 9 Mar 2022 • Andong Deng, Xingjian Li, Di Hu, Tianyang Wang, Haoyi Xiong, Chengzhong Xu
Based on the contradictory phenomenon between FE and FT that better feature extractor fails to be fine-tuned better accordingly, we conduct comprehensive analyses on features before softmax layer to provide insightful explanations.
1 code implementation • 10 Dec 2021 • Tianyang Wang, Xingjian Li, Pengkun Yang, Guosheng Hu, Xiangrui Zeng, Siyu Huang, Cheng-Zhong Xu, Min Xu
In this work, we explore such an impact by theoretically proving that selecting unlabeled data of higher gradient norm leads to a lower upper-bound of test loss, resulting in better test performance.
no code implementations • NAACL 2021 • Hang Hua, Xingjian Li, Dejing Dou, Cheng-Zhong Xu, Jiebo Luo
The brittleness of this process is often reflected by the sensitivity to random seeds.
no code implementations • 25 Mar 2021 • Xingjian Li, Haoyi Xiong, Chengzhong Xu, Dejing Dou
Performing mixup for transfer learning with pre-trained models however is not that simple, a high capacity pre-trained model with a large fully-connected (FC) layer could easily overfit to the target dataset even with samples-to-labels mixed up.
1 code implementation • 19 Mar 2021 • Xuhong LI, Haoyi Xiong, Xingjian Li, Xuanyu Wu, Xiao Zhang, Ji Liu, Jiang Bian, Dejing Dou
Then, to understand the interpretation results, we also survey the performance metrics for evaluating interpretation algorithms.
1 code implementation • CVPR 2021 • Abulikemu Abuduweili, Xingjian Li, Humphrey Shi, Cheng-Zhong Xu, Dejing Dou
To better exploit the value of both pre-trained weights and unlabeled target examples, we introduce adaptive consistency regularization that consists of two complementary components: Adaptive Knowledge Consistency (AKC) on the examples between the source and target model, and Adaptive Representation Consistency (ARC) on the target model between labeled and unlabeled examples.
no code implementations • 1 Jan 2021 • Xiao Zhang, Di Hu, Xingjian Li, Dejing Dou, Ji Wu
We demonstrate using model information as a general analysis tool to gain insight into problems that arise in deep learning.
no code implementations • 1 Jan 2021 • Haozhe An, Haoyi Xiong, Xuhong LI, Xingjian Li, Dejing Dou, Zhanxing Zhu
The recent theoretical investigation (Li et al., 2020) on the upper bound of generalization error of deep neural networks (DNNs) demonstrates the potential of using the gradient norm as a measure that complements validation accuracy for model selection in practice.
1 code implementation • 14 Dec 2020 • Dong Wang, Di Hu, Xingjian Li, Dejing Dou
The main reason is that large number of nodes (i. e., video frames) makes GCNs hard to capture and model temporal relations in videos.
Ranked #22 on
Action Segmentation
on Breakfast
1 code implementation • 9 Nov 2020 • Derek Onken, Levon Nurbekyan, Xingjian Li, Samy Wu Fung, Stanley Osher, Lars Ruthotto
Our approach is grid-free and scales efficiently to dimensions where grids become impractical or infeasible.
Optimization and Control
no code implementations • 16 Oct 2020 • Xingjian Li, Di Hu, Xuhong LI, Haoyi Xiong, Zhi Ye, Zhipeng Wang, Chengzhong Xu, Dejing Dou
Fine-tuning deep neural networks pre-trained on large scale datasets is one of the most practical transfer learning paradigm given limited quantity of training samples.
no code implementations • 16 Sep 2020 • Xiao Zhang, Xingjian Li, Dejing Dou, Ji Wu
We propose a practical measure of the generalizable information in a neural network model based on prequential coding, which we term Information Transfer ($L_{IT}$).
no code implementations • 20 Jul 2020 • Xingjian Li, Haoyi Xiong, Haozhe An, Cheng-Zhong Xu, Dejing Dou
While the existing multitask learning algorithms need to run backpropagation over both the source and target datasets and usually consume a higher gradient complexity, XMixup transfers the knowledge from source to target tasks more efficiently: for every class of the target task, XMixup selects the auxiliary samples from the source dataset and augments training samples via the simple mixup strategy.
1 code implementation • ICML 2020 • Xingjian Li, Haoyi Xiong, Haozhe An, Cheng-Zhong Xu, Dejing Dou
RIFLE brings meaningful updates to the weights of deep CNN layers and improves low-level feature learning, while the effects of randomization can be easily converged throughout the overall learning procedure.
1 code implementation • 29 May 2020 • Derek Onken, Samy Wu Fung, Xingjian Li, Lars Ruthotto
On five high-dimensional density estimation and generative modeling tasks, OT-Flow performs competitively to state-of-the-art CNFs while on average requiring one-fourth of the number of weights with an 8x speedup in training time and 24x speedup in inference.
no code implementations • ICLR 2020 • Kafeng Wang, Xitong Gao, Yiren Zhao, Xingjian Li, Dejing Dou, Cheng-Zhong Xu
Deep convolutional neural networks are now widely deployed in vision applications, but a limited size of training data can restrict their task performance.
no code implementations • 26 Apr 2020 • Xingjian Li, Haoyi Xiong, Haozhe An, Dejing Dou, Chengzhong Xu
Softening labels of training datasets with respect to data representations has been frequently used to improve the training of deep neural networks (DNNs).
no code implementations • 18 Nov 2019 • Ruosi Wan, Haoyi Xiong, Xingjian Li, Zhanxing Zhu, Jun Huan
The empirical results show that the proposed descent direction estimation strategy DTNH can always improve the performance of deep transfer learning tasks based on all above regularizers, even when transferring pre-trained weights from inappropriate networks.
no code implementations • 23 Aug 2019 • Dou Goodman, Xingjian Li, Ji Liu, Dejing Dou, Tao Wei
Finally, we conduct extensive experiments using a wide range of datasets and the experiment results show that our AT+ALP achieves the state of the art defense performance.
no code implementations • 3 Feb 2019 • Yingzhen Yang, Jiahui Yu, Xingjian Li, Jun Huan, Thomas S. Huang
In this paper, we investigate the role of Rademacher complexity in improving generalization of DNNs and propose a novel regularizer rooted in Local Rademacher Complexity (LRC).
2 code implementations • ICLR 2019 • Xingjian Li, Haoyi Xiong, Hanchao Wang, Yuxuan Rao, Li-Ping Liu, Zeyu Chen, Jun Huan
Instead of constraining the weights of neural network, DELTA aims to preserve the outer layer outputs of the target network.