1 code implementation • 24 Feb 2025 • Huanghai Liu, Quzhe Huang, Qingjing Chen, Yiran Hu, Jiayu Ma, Yun Liu, Weixing Shen, Yansong Feng
The Four-Element Theory is a fundamental framework in criminal law, defining the constitution of crime through four dimensions: Subject, Object, Subjective aspect, and Objective aspect.
no code implementations • 3 Jan 2025 • Kangcheng Luo, Quzhe Huang, Cong Jiang, Yansong Feng
Multi-faceted evaluations by legal experts indicate that the quality of our concept interpretations is comparable to those written by human experts.
1 code implementation • 8 Oct 2024 • Yang Jin, Zhicheng Sun, Ningyuan Li, Kun Xu, Hao Jiang, Nan Zhuang, Quzhe Huang, Yang song, Yadong Mu, Zhouchen Lin
Video generation requires modeling a vast spatiotemporal space, which demands significant computational resources and data usage.
no code implementations • 14 Aug 2024 • Yutong Hu, Quzhe Huang, Yansong Feng
Event Temporal Relation Extraction (ETRE) aims to identify the temporal relationship between two events, which plays an important role in natural language understanding.
no code implementations • 4 Jul 2024 • Mingxu Tao, Chen Zhang, Quzhe Huang, Tianyao Ma, Songfang Huang, Dongyan Zhao, Yansong Feng
Adapting large language models (LLMs) to new languages typically involves continual pre-training (CT) followed by supervised fine-tuning (SFT).
no code implementations • 17 Jun 2024 • Yutong Hu, Quzhe Huang, Kangcheng Luo, Yansong Feng
In this paper, we aim to explore which kinds of words benefit more from long contexts in language models.
no code implementations • 9 May 2024 • Yutong Hu, Quzhe Huang, Mingxu Tao, Chen Zhang, Yansong Feng
Recent studies have shown that Large Language Models (LLMs) have the potential to process extremely long text.
1 code implementation • 12 Mar 2024 • Quzhe Huang, Zhenwei An, Nan Zhuang, Mingxu Tao, Chen Zhang, Yang Jin, Kun Xu, Liwei Chen, Songfang Huang, Yansong Feng
In this paper, we introduce a novel dynamic expert selection framework for Mixture of Experts (MoE) models, aiming to enhance computational efficiency and model performance by adjusting the number of activated experts based on input difficulty.
1 code implementation • 27 Feb 2024 • Mingxu Tao, Quzhe Huang, Kun Xu, Liwei Chen, Yansong Feng, Dongyan Zhao
The advancement of Multimodal Large Language Models (MLLMs) has greatly accelerated the development of applications in understanding integrated texts and images.
1 code implementation • 5 Feb 2024 • Yang Jin, Zhicheng Sun, Kun Xu, Liwei Chen, Hao Jiang, Quzhe Huang, Chengru Song, Yuliang Liu, Di Zhang, Yang song, Kun Gai, Yadong Mu
In light of recent advances in multimodal Large Language Models (LLMs), there is increasing attention to scaling them from image-text data to more informative real-world videos.
Ranked #3 on
Text-to-Video Generation
on MSR-VTT
1 code implementation • 19 Dec 2023 • Haowei Du, Quzhe Huang, Chen Li, Chen Zhang, Yang Li, Dongyan Zhao
To address this issue, we construct a \textbf{dual relation graph} where each node denotes a relation in the original KG (\textbf{primal entity graph}) and edges are constructed between relations sharing same head or tail entities.
2 code implementations • 14 Nov 2023 • Chen Zhang, Mingxu Tao, Quzhe Huang, Jiuheng Lin, Zhibin Chen, Yansong Feng
To address this accessibility challenge, we present MC$^2$, a Multilingual Corpus of Minority Languages in China, which is the largest open-source corpus of its kind so far.
1 code implementation • 25 Oct 2023 • Quzhe Huang, Yanxi Zhang, Dongyan Zhao
These methods extract events according to their appearance order in the document, however, the event that appears in the first sentence does not mean that it is the easiest to extract.
1 code implementation • 9 Sep 2023 • Yang Jin, Kun Xu, Liwei Chen, Chao Liao, Jianchao Tan, Quzhe Huang, Bin Chen, Chenyi Lei, An Liu, Chengru Song, Xiaoqiang Lei, Di Zhang, Wenwu Ou, Kun Gai, Yadong Mu
Specifically, we introduce a well-designed visual tokenizer to translate the non-linguistic image into a sequence of discrete tokens like a foreign language that LLM can read.
no code implementations • 28 May 2023 • Quzhe Huang, Yutong Hu, Shengqi Zhu, Yansong Feng, Chang Liu, Dongyan Zhao
After examining the relation definitions in various ETRE tasks, we observe that all relations can be interpreted using the start and end time points of events.
1 code implementation • 24 May 2023 • Quzhe Huang, Mingxu Tao, Chen Zhang, Zhenwei An, Cong Jiang, Zhibin Chen, Zirui Wu, Yansong Feng
Specifically, we inject domain knowledge during the continual training stage and teach the model to learn professional skills using properly designed supervised fine-tuning tasks.
1 code implementation • 31 Oct 2022 • Zhenwei An, Quzhe Huang, Cong Jiang, Yansong Feng, Dongyan Zhao
The charge prediction task aims to predict the charge for a case given its fact description.
no code implementations • 7 Sep 2022 • Haowei Du, Quzhe Huang, Chen Zhang, Dongyan Zhao
Multi-hop Knowledge Base Question Answering(KBQA) aims to find the answer entity in a knowledge base which is several hops from the topic entity mentioned in the question.
1 code implementation • ACL 2022 • Quzhe Huang, Shibo Hao, Yuan Ye, Shengqi Zhu, Yansong Feng, Dongyan Zhao
DocRED is a widely used dataset for document-level relation extraction.
no code implementations • ACL 2021 • Quzhe Huang, Shengqi Zhu, Yansong Feng, Dongyan Zhao
Recent studies strive to incorporate various human rationales into neural networks to improve model performance, but few pay attention to the quality of the rationales.
1 code implementation • ACL 2021 • Quzhe Huang, Shengqi Zhu, Yansong Feng, Yuan Ye, Yuxuan Lai, Dongyan Zhao
Document-level Relation Extraction (RE) is a more challenging task than sentence RE as it often requires reasoning over multiple sentences.
Ranked #48 on
Relation Extraction
on DocRED
1 code implementation • Findings (ACL) 2021 • Yuxuan Lai, Chen Zhang, Yansong Feng, Quzhe Huang, Dongyan Zhao
A thorough empirical analysis shows that MRC models tend to learn shortcut questions earlier than challenging questions, and the high proportions of shortcut questions in training sets hinder models from exploring the sophisticated reasoning skills in the later stage of training.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Xiaohan Yu, Quzhe Huang, Zheng Wang, Yansong Feng, Dongyan Zhao
Code comments are vital for software maintenance and comprehension, but many software projects suffer from the lack of meaningful and up-to-date comments in practice.