no code implementations • NAACL (SUKI) 2022 • Xiaofeng Chen, YiRong Chen, Xiaofen Xing, Xiangmin Xu, Wenjing Han, Qianfeng Tie
Because of the compositionality of natural language, syntactic structure which contains the information about the relationship between words is a key factor for semantic understanding.
no code implementations • 7 Jul 2024 • Yan Wang, Yawen Zeng, Jingsheng Zheng, Xiaofen Xing, Jin Xu, Xiangmin Xu
Therefore, we try to explore the collection of CoT datasets in videos to lead to video OpenQA and improve the reasoning ability of MLLMs.
1 code implementation • 3 Jul 2024 • Rui Liu, Haolin Zuo, Zheng Lian, Xiaofen Xing, Björn W. Schuller, Haizhou Li
Together with the release of the dataset, we also develop an Emotion and Intent Interaction (EI$^2$) network as a reference system by modeling the deep correlation between emotion and intent in the multimodal conversation.
no code implementations • 12 Mar 2024 • Weiying Xue, Qi Liu, Qiwei Xiong, Yuxiao Wang, Zhenao Wei, Xiaofen Xing, Xiangmin Xu
Human-object interaction (HOI) detection aims to locate human-object pairs and identify their interaction categories in images.
1 code implementation • 5 Mar 2024 • Xiangyu Li, Xinjie Shen, Yawen Zeng, Xiaofen Xing, Jin Xu
However, compared with financial institutions, it is not easy for ordinary investors to mine factors and analyze news.
Ranked #1 on Stock Market Prediction on Astock
1 code implementation • 4 Mar 2024 • Baozhu Zhao, Qiwei Xiong, Xiaohan Zhang, Jingfeng Guo, Qi Liu, Xiaofen Xing, Xiangmin Xu
Three-dimensional point cloud anomaly detection that aims to detect anomaly data points from a training set serves as the foundation for a variety of applications, including industrial inspection and autonomous driving.
no code implementations • 1 Nov 2023 • YiRong Chen, Xiaofen Xing, Jingkai Lin, huimin zheng, Zhenyu Wang, Qi Liu, Xiangmin Xu
Large language models (LLMs) have been widely applied in various fields due to their excellent capability for memorizing knowledge and chain of thought (CoT).
1 code implementation • 24 Oct 2023 • YiRong Chen, Zhenyu Wang, Xiaofen Xing, huimin zheng, Zhipei Xu, Kai Fang, Junhong Wang, Sihang Li, Jieling Wu, Qi Liu, Xiangmin Xu
Large language models (LLMs) have performed well in providing general and extensive health suggestions in single-turn conversations, exemplified by systems such as ChatGPT, ChatGLM, ChatDoctor, DoctorGLM, and etc.
no code implementations • 17 Oct 2023 • Zhaojie Chu, Kailing Guo, Xiaofen Xing, Yilin Lan, Bolun Cai, Xiangmin Xu
In this study, we propose a novel framework, CorrTalk, which effectively establishes the temporal correlation between hierarchical speech features and facial activities of different intensities across distinct regions.
no code implementations • 25 Sep 2023 • Pucheng Zhai, Kailing Guo, Fang Liu, Xiaofen Xing, Xiangmin Xu
Therefore the pruning strategy can gradually prune the network and automatically determine the appropriate pruning rates for each layer.
1 code implementation • 20 Jul 2023 • Weidong Chen, Xiaofen Xing, Peihao Chen, Xiangmin Xu
Although PTMs shed new light on artificial general intelligence, they are constructed with general tasks in mind, and thus, their efficacy for specific tasks can be further improved.
1 code implementation • 3 Mar 2023 • Shuaiqi Chen, Xiaofen Xing, Weibin Zhang, Weidong Chen, Xiangmin Xu
Self-attention mechanism is applied within windows for capturing temporal important information locally in a fine-grained way.
1 code implementation • 27 Feb 2023 • Weidong Chen, Xiaofen Xing, Xiangmin Xu, Jianxin Pang, Lan Du
Paralinguistic speech processing is important in addressing many issues, such as sentiment and neurocognitive disorder analyses.
1 code implementation • 29 May 2022 • YiRong Chen, Weiquan Fan, Xiaofen Xing, Jianxin Pang, Minlie Huang, Wenjing Han, Qianfeng Tie, Xiangmin Xu
Finally, we provide baseline systems for these tasks and consider the function of speakers' personalities and emotions on conversation.
Ranked #1 on Emotion Recognition in Conversation on CPED
1 code implementation • 12 Apr 2022 • Kailing Guo, Zhenquan Lin, Canyang Chen, Xiaofen Xing, Fang Liu, Xiangmin Xu
We theoretically show that energy transfer eases the trend of gradient vanishing caused by projection.
1 code implementation • 9 Oct 2021 • Zhenquan Lin, Kailing Guo, Xiaofen Xing, Xiangmin Xu
Comprehensive experiments show that WE outperforms the other reactivation methods and plug-in training methods with typical convolutional neural networks, especially lightweight networks.
1 code implementation • 30 Jan 2021 • Weiquan Fan, Xiangmin Xu, Xiaofen Xing, Weidong Chen, DongYan Huang
Speech emotion recognition is a vital contributor to the next generation of human-computer interaction (HCI).
Ranked #1 on Speech Emotion Recognition on LSSED
1 code implementation • 14 May 2019 • Weirui Lu, Xiaofen Xing, Bolun Cai, Xiangmin Xu
However, the performance of ranking-based methods is often poor and this is mainly due to two reasons: 1) image cropping is a listwise ranking task rather than pairwise comparison; 2) the rescaling caused by pooling layer and the deformation in view generation damage the performance of composition learning.
1 code implementation • 23 Apr 2019 • Bolun Cai, Xiangmin Xu, Xiaofen Xing, Kui Jia, Jie Miao, DaCheng Tao
Visual tracking is challenging due to image variations caused by various factors, such as object deformation, scale change, illumination change and occlusion.
no code implementations • 9 Jun 2017 • Jie Miao, Xiangmin Xu, Xiaofen Xing, DaCheng Tao
However, complex temporal variations require high-level semantic representations to fully achieve temporal slowness, and thus it is impractical to learn a high-level representation from dynamic textures directly by SFA.