1 code implementation • 30 May 2024 • Chong Li, Wen Yang, Jiajun Zhang, Jinliang Lu, Shaonan Wang, Chengqing Zong
In addition, we find that models tuned on cross-lingual instruction following samples can follow the instruction in the output language without further tuning.
no code implementations • 16 Apr 2024 • Sicheng Xu, Guojun Chen, Yu-Xiao Guo, Jiaolong Yang, Chong Li, Zhenyu Zang, Yizhong Zhang, Xin Tong, Baining Guo
We introduce VASA, a framework for generating lifelike talking faces with appealing visual affective skills (VAS) given a single static image and a speech audio clip.
no code implementations • 27 Mar 2024 • Jingyang Huo, Yikai Wang, Xuelin Qian, Yun Wang, Chong Li, Jianfeng Feng, Yanwei Fu
Recent fMRI-to-image approaches mainly focused on associating fMRI signals with specific conditions of pre-trained diffusion models.
no code implementations • 2 Mar 2024 • Yunhao Zhang, Xiaohan Zhang, Chong Li, Shaonan Wang, Chengqing Zong
Results show that language models share significant similarities with human cognitive data and the similarity patterns are modulated by the data modality and stimuli complexity.
no code implementations • 20 Nov 2023 • Chong Li, Zhun Fan, Ying Chen, Huibiao Lin, Laura Moretti, Giuseppe Loprencipe, Weihua Sheng, Kelvin C. P. Wang
Meanwhile, these models can not automatically correct errors in the prediction, nor can it adapt to the changes of the environment to automatically extract and detect thin cracks.
1 code implementation • 14 Nov 2023 • Chong Li, Shaonan Wang, Jiajun Zhang, Chengqing Zong
It aligns the internal sentence representations across different languages via multilingual contrastive learning and aligns outputs by following cross-lingual instructions in the target language.
1 code implementation • 16 Oct 2023 • Chong Li, Shaonan Wang, Yunhao Zhang, Jiajun Zhang, Chengqing Zong
We further propose a simple multi-task training method to increase functional specialization and mitigate negative information transfer in multi-task learning.
2 code implementations • 29 May 2023 • Wen Yang, Chong Li, Jiajun Zhang, Chengqing Zong
Second, we continue training the model with a large-scale parallel dataset that covers 102 natural languages.
Ranked #4 on Machine Translation on FLoRes-200
no code implementations • 28 Feb 2023 • Yizhong Zhang, Zhiqi Li, Sicheng Xu, Chong Li, Jiaolong Yang, Xin Tong, Baining Guo
A key challenge in emulating the remote hand touch is the realistic rendering of the participant's hand and arm as the hand touches the screen.
no code implementations • 19 Dec 2022 • Yangyu Huang, Xi Chen, Jongyoo Kim, Hao Yang, Chong Li, Jiaolong Yang, Dong Chen
To evaluate our method, we manually label the dense landmarks on 300W testset.
Ranked #1 on Face Alignment on 300W
no code implementations • 16 Dec 2022 • Yuanjie Xia, Mengyao Yuan, Alexandra Dobrea, Chong Li, Hadi Heidari, Nigel Mottram, Rami Ghannam
The antenna is one of the key building blocks of many wearable electronic device, and its functions include wireless communications, energy harvesting and radiative wireless power transfer (WPT).
1 code implementation • CVPR 2022 • Minghao Chen, Fangyun Wei, Chong Li, Deng Cai
In this paper, we introduce a novel contrastive action representation learning (CARL) framework to learn frame-wise action representations, especially for long videos, in a self-supervised manner.
no code implementations • 29 Sep 2021 • Quentin Petit, Chong Li, Kelun Chai, Serge G Petiton
Introducing random edge dropping from the input graph at training epochs could reduce over-fitting and over-smoothing phenomenon and increase the depth of GNNs.
1 code implementation • ICCV 2021 • Yangyu Huang, Hao Yang, Chong Li, Jongyoo Kim, Fangyun Wei
On the other hand, AAM is an attention module which can get anisotropic attention mask focusing on the region of point and its local edge connected by adjacent points, it has a stronger response in tangent than in normal, which means relaxed constraints in the tangent.
Ranked #5 on Face Alignment on 300W
2 code implementations • Part of the Lecture Notes in Computer Science book series 2021 • Haoran Wang, Chong Li, Thibaut Tachon, Hongxing Wang, Sheng Yang, Sébastien Limet, Sophie Robert
We propose the Flex-Edge Recursive Graph and the Double Recursive Algorithm, successfully limiting our parallelization strategy generation to a linear complexity with a good quality of parallelization strategy.
1 code implementation • ACL 2021 • Chong Li, Cenyuan Zhang, Xiaoqing Zheng, Xuanjing Huang
A sequence-to-sequence learning with neural networks has empirically proven to be an effective framework for Chinese Spelling Correction (CSC), which takes a sentence with some spelling errors as input and outputs the corrected one.
no code implementations • 24 Feb 2021 • Yuehui Ma, Hongchi Wang, Chong Li, Lianghao Lin, Yan Sun, Ji Yang
About 78\% of the N-PDFs of the selected molecular clouds are well fitted with log-normal functions with only small deviations at high-densities which correspond to star-forming regions with scales of $\sim$1-5 pc in the Local arm and $\sim$5-10 pc in the Perseus arm.
Astrophysics of Galaxies
no code implementations • 1 Jul 2020 • Zhun Fan, Chong Li, Ying Chen, Jiahong Wei, Giuseppe Loprencipe, Xiaopeng Chen, Paola Di Mascio
Finally, the hierarchical feature learning module is designed to obtain a multi-scale features from the high to low-level convolutional layers, which are integrated to predict pixel-wise crack detection.
no code implementations • 8 Feb 2020 • Zhun Fan, Chong Li, Ying Chen, Paola Di Mascio, Xiaopeng Chen, Guijie Zhu, Giuseppe Loprencipe
In this paper, we propose an ensemble of convolutional neural networks (without a pooling layer) based on probability fusion for automated pavement crack detection and measurement.
no code implementations • 12 Nov 2019 • Xin Li, Yaohua Hu, Chong Li, Xiaoqi Yang, Tianzi Jiang
In this paper, we discuss the statistical properties of the $\ell_q$ optimization methods $(0<q\leq 1)$, including the $\ell_q$ minimization method and the $\ell_q$ regularization method, for estimating a sparse parameter from noisy observations in high-dimensional linear regression with either a deterministic or random design.
no code implementations • 25 Sep 2019 • Chong Li, Dan Shen, C.J. Richard Shi, Hongxia Yang
We propose a novel method to estimate the parameters of a collection of Hidden Markov Models (HMM), each of which corresponds to a set of known features.
2 code implementations • 28 Jun 2019 • Zhun Fan, Jiajie Mo, Benzhang Qiu, Wenji Li, Guijie Zhu, Chong Li, Jianye Hu, Yibiao Rong, Xinjian Chen
Compared with other convolution networks utilizing standard convolution for feature extraction, the proposed method utilizes octave convolutions and octave transposed convolutions for learning multiple-spatial-frequency features, thus can better capture retinal vasculatures with varying sizes and shapes.
1 code implementation • 14 Sep 2018 • Mehmet Necip Kurt, Oyetunji Ogundijo, Chong Li, Xiaodong Wang
Early detection of cyber-attacks is crucial for a safe and reliable operation of the smart grid.
1 code implementation • ECCV 2018 • Chong Li, C. J. Richard Shi
We present COBLA---Constrained Optimization Based Low-rank Approximation---a systematic method of finding an optimal low-rank approximation of a trained convolutional neural network, subject to constraints in the number of multiply-accumulate (MAC) operations and the memory footprint.
no code implementations • 3 Feb 2018 • Himani Arora, Rajath Kumar, Jason Krone, Chong Li
In addition, most reinforcement learning research on multi-task learning has been focused on discrete action spaces, which are not used for robotic control in the real-world.