1 code implementation • 27 May 2025 • Xuanle Zhao, Zilin Sang, YuXuan Li, Qi Shi, Weilun Zhao, Shuo Wang, Duzhen Zhang, Xu Han, Zhiyuan Liu, Maosong Sun
Building on this idea, we propose AutoReproduce, a multi-agent framework capable of automatically reproducing experiments described in research papers in an end-to-end manner.
1 code implementation • 17 May 2025 • Junhao Zheng, Xidi Cai, Qiuke Li, Duzhen Zhang, Zhongzhi Li, Yingying Zhang, Le Song, Qianli Ma
Lifelong learning is essential for intelligent agents operating in dynamic environments.
no code implementations • 24 Apr 2025 • Cheng Wang, Yue Liu, Baolong Bi, Duzhen Zhang, Zhongzhi Li, Junfeng Fang, Bryan Hooi
Large Reasoning Models (LRMs) have exhibited extraordinary prowess in tasks like mathematics and coding, leveraging their advanced reasoning capabilities.
1 code implementation • 24 Feb 2025 • Zhong-Zhi Li, Duzhen Zhang, Ming-Liang Zhang, Jiaxin Zhang, Zengyan Liu, Yuxuan Yao, Haotian Xu, Junhao Zheng, Pei-Jie Wang, Xiuyi Chen, Yingying Zhang, Fei Yin, Jiahua Dong, Zhijiang Guo, Le Song, Cheng-Lin Liu
Achieving human-level intelligence requires refining the transition from the fast, intuitive System 1 to the slower, more deliberate System 2 reasoning.
no code implementations • 18 Feb 2025 • Zixiao Wang, Duzhen Zhang, Ishita Agrawal, Shen Gao, Le Song, Xiuying Chen
These include a pre-training task focused on mastering external linguistic structures and knowledge, as well as three fine-tuning tasks: multiple-choice question answering, generative question answering, and style transfer, each aligning the LLM with Lu Xun's internal ideation and writing style.
1 code implementation • 13 Jan 2025 • Junhao Zheng, Chengming Shi, Xidi Cai, Qiuke Li, Duzhen Zhang, Chenxing Li, Dong Yu, Qianli Ma
This survey is the first to systematically summarize the potential techniques for incorporating lifelong learning into LLM-based agents.
1 code implementation • 18 Nov 2024 • Duzhen Zhang, Yahan Yu, Chenxing Li, Jiahua Dong, Dong Yu
In a more realistic scenario, local clients receive new entity types continuously, while new local clients collecting novel data may irregularly join the global FNER training.
1 code implementation • 23 Oct 2024 • Jiahua Dong, Wenqi Liang, Hongliu Li, Duzhen Zhang, Meng Cao, Henghui Ding, Salman Khan, Fahad Shahbaz Khan
Moreover, they heavily suffer from catastrophic forgetting and concept neglect on old personalized concepts when continually learning a series of new concepts.
no code implementations • 11 May 2024 • Manjie Xu, Chenxing Li, Duzhen Zhang, Dan Su, Wei Liang, Dong Yu
Audio editing involves the arbitrary manipulation of audio content through precise control.
no code implementations • 29 Mar 2024 • Duzhen Zhang, Qingyu Wang, Tielin Zhang, Bo Xu
Diverging from the conventional direct linear weighted sum, the BPT-SAN models the local nonlinearities of dendritic trees within the inter-layer connections.
no code implementations • 27 Mar 2024 • Qingyu Wang, Duzhen Zhang, Tilelin Zhang, Bo Xu
Energy-efficient spikformer has been proposed by integrating the biologically plausible spiking neural network (SNN) and artificial Transformer, whereby the Spiking Self-Attention (SSA) is used to achieve both higher accuracy and lower computational cost.
no code implementations • 24 Jan 2024 • Duzhen Zhang, Yahan Yu, Jiahua Dong, Chenxing Li, Dan Su, Chenhui Chu, Dong Yu
In the past year, MultiModal Large Language Models (MM-LLMs) have undergone substantial advancements, augmenting off-the-shelf LLMs to support MM inputs or outputs via cost-effective training strategies.
1 code implementation • 23 Oct 2023 • Duzhen Zhang, Wei Cong, Jiahua Dong, Yahan Yu, Xiuyi Chen, Yonggang Zhang, Zhen Fang
This issue is intensified in CNER due to the consolidation of old entity types from previous steps into the non-entity type at each step, leading to what is known as the semantic shift problem of the non-entity type.
Continual Named Entity Recognition
named-entity-recognition
+1
1 code implementation • 17 Aug 2023 • Duzhen Zhang, Hongliu Li, Wei Cong, Rongtao Xu, Jiahua Dong, Xiuyi Chen
However, INER faces the challenge of catastrophic forgetting specific for incremental learning, further aggravated by background shift (i. e., old and future entity types are labeled as the non-entity type in the current task).
no code implementations • 2 Aug 2023 • Qingyu Wang, Duzhen Zhang, Tielin Zhang, Bo Xu
The results indicate that compared to the SOTA Spikformer with SSA, Spikformer with LT achieves higher Top-1 accuracy on neuromorphic datasets (i. e., CIFAR10-DVS and DVS128 Gesture) and comparable Top-1 accuracy on static datasets (i. e., CIFAR-10 and CIFAR-100).
1 code implementation • CVPR 2023 • Jiahua Dong, Duzhen Zhang, Yang Cong, Wei Cong, Henghui Ding, Dengxin Dai
Moreover, new clients collecting novel classes may join in the global training of FSS, which further exacerbates catastrophic forgetting.
no code implementations • 3 Feb 2023 • Yumin Zhang, Yajun Gao, Hongliu Li, Ating Yin, Duzhen Zhang, Xiuyi Chen
Unsupervised Domain Adaptation (UDA), which aims to explore the transferrable features from a well-labeled source domain to a related unlabeled target domain, has been widely progressed.
1 code implementation • 2 Feb 2023 • Minglun Han, Qingyu Wang, Tielin Zhang, Yi Wang, Duzhen Zhang, Bo Xu
The spiking neural network (SNN) using leaky-integrated-and-fire (LIF) neurons has been commonly used in automatic speech recognition (ASR) tasks.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+1
1 code implementation • 29 Dec 2022 • Duzhen Zhang, Tielin Zhang, Shuncheng Jia, Qingyu Wang, Bo Xu
Learning from interaction is the primary way that biological agents acquire knowledge about their environment and themselves.
no code implementations • 24 May 2022 • Feilong Chen, Xiuyi Chen, Jiaxin Shi, Duzhen Zhang, Jianlong Chang, Qi Tian
It also achieves about +4. 9 AR on COCO and +3. 8 AR on Flickr30K than LightingDot and achieves comparable performance with the state-of-the-art (SOTA) fusion-based model METER.
1 code implementation • 12 Mar 2022 • Duzhen Zhang, Shuncheng Jia, Qingyu Wang
In recent years, spiking neural networks (SNNs) have received extensive attention in brain-inspired intelligence due to their rich spatially-temporal dynamics, various encoding methods, and event-driven characteristics that naturally fit the neuromorphic hardware.
3 code implementations • COLING 2022 • Duzhen Zhang, Zhen Yang, Fandong Meng, Xiuyi Chen, Jie zhou
Causal Emotion Entailment (CEE) aims to discover the potential causes behind an emotion in a conversational utterance.
Ranked #4 on
Causal Emotion Entailment
on RECCON
1 code implementation • 18 Feb 2022 • Feilong Chen, Duzhen Zhang, Minglun Han, Xiuyi Chen, Jing Shi, Shuang Xu, Bo Xu
Finally, we discuss the new frontiers in VLP.
no code implementations • 15 Jun 2021 • Duzhen Zhang, Tielin Zhang, Shuncheng Jia, Xiang Cheng, Bo Xu
Based on a hybrid learning framework, where a spike actor-network infers actions from states and a deep critic network evaluates the actor, we propose a Population-coding and Dynamic-neurons improved Spiking Actor Network (PDSAN) for efficient state representation from two different scales: input coding and neuronal coding.
no code implementations • COLING 2020 • Duzhen Zhang, Xiuyi Chen, Shuang Xu, Bo Xu
For one thing, speakers often rely on the context and commonsense knowledge to express emotions; for another, most utterances contain neutral emotion in conversations, as a result, the confusion between a few non-neutral utterances and much more neutral ones restrains the emotion recognition performance.