1 code implementation • COLING 2022 • Hongxiao Zhang, Hui Huang, Jiale Gao, Yufeng Chen, Jinan Xu, Jian Liu
In this paper, we propose an Iterative Constrained Back-Translation (ICBT) method to incorporate in-domain lexical knowledge on the basis of BT for unsupervised domain adaptation of NMT.
no code implementations • Findings (EMNLP) 2021 • Siyu Lai, Hui Huang, Dong Jing, Yufeng Chen, Jinan Xu, Jian Liu
Recent multilingual pre-trained models, like XLM-RoBERTa (XLM-R), have been demonstrated effective in many cross-lingual tasks.
Cross-Lingual Sentiment Classification Dialogue State Tracking +4
no code implementations • CCL 2020 • Xingchen Li, Mingtong Liu, Yujie Zhang, Jinan Xu, Yufeng Chen
The experimental results on the Penn Chinese treebank (CTB5) show that our proposed joint model improved by 0. 38% on dependency parsing than the model of Yan et al. (2019).
no code implementations • CCL 2021 • Shuning Lv, Jian Liu, Jinan Xu, Yufeng Chen, Yujie Zhang
“实体边界预测对中文命名实体识别至关重要。现有研究为改善边界识别效果提出的多任务学习方法仅考虑与分词任务结合, 缺少多任务标签训练数据, 无法学到任务的标签一致性关系。本文提出一种新的基于多任务标签一致性机制的中文命名实体识别方法:将分词和词性信息融入命名实体识别模型, 使三种任务联合训练;建立基于标签一致性机制的多任务学习模式, 来捕获标签一致性关系及学习多任务表示。全样本和小样本实验表明了方法的有效性。”
Chinese Named Entity Recognition named-entity-recognition +1
1 code implementation • ACL 2022 • Jian Liu, Yufeng Chen, Jinan Xu
Event detection (ED) is a critical subtask of event extraction that seeks to identify event triggers of certain types in texts. Despite significant advances in ED, existing methods typically follow a “one model fits all types” approach, which sees no differences between event types and often results in a quite skewed performance. Finding the causes of skewed performance is crucial for the robustness of an ED model, but to date there has been little exploration of this problem. This research examines the issue in depth and presents a new concept termed trigger salience attribution, which can explicitly quantify the underlying patterns of events.
no code implementations • CCL 2021 • Bo Jin, Mingtong Liu, Yujie Zhang, Jinan Xu, Yufeng Chen
“如何挖掘语言资源中丰富的复述模板, 是复述研究中的一项重要任务。已有方法在人工给定种子实体对的基础上, 利用实体关系, 通过自举迭代方式, 从开放域获取复述模板, 规避对平行语料或可比语料的依赖, 但是该方法需人工给定实体对, 实体关系受限;在迭代过程中语义会发生偏移, 影响获取质量。针对这些问题, 我们考虑知识库中包含描述特定语义关系的实体对(即关系三元组), 提出融合外部知识的开放域复述模板自动获取方法。首先, 将关系三元组与开放域文本对齐, 获取关系对应文本, 并将文本中语义丰富部分泛化成变量槽, 获取关系模板;接着设计模板表示方法, 本文利用预训练语言模型, 在模板表示中融合变量槽语义;最后, 根据获得的模板表示, 设计自动聚类与筛选方法, 获取高精度的复述模板。在融合自动评测与人工评测的评价方法下, 实验结果表明, 本文提出的方法实现了在开放域数据上复述模板的自动泛化与获取, 能够获得质量高、语义一致的复述模板。”
1 code implementation • EMNLP 2021 • Erguang Yang, Mingtong Liu, Deyi Xiong, Yujie Zhang, Yao Meng, Changjian Hu, Jinan Xu, Yufeng Chen
Particularly, we design a two-stage learning method to effectively train the model using non-parallel data.
no code implementations • Findings (NAACL) 2022 • Erguang Yang, Chenglin Bai, Deyi Xiong, Yujie Zhang, Yao Meng, Jinan Xu, Yufeng Chen
To model the alignment relation between words and nodes, we propose an attention regularization objective, which makes the decoder accurately select corresponding syntax nodes to guide the generation of words. Experiments show that SI-SCP achieves state-of-the-art performances in terms of semantic and syntactic quality on two popular benchmark datasets. Additionally, we propose a Syntactic Template Retriever (STR) to retrieve compatible syntactic structures.
no code implementations • EMNLP 2021 • Jian Liu, Yufeng Chen, Jinan Xu
Implicit event argument extraction (EAE) is a crucial document-level information extraction task that aims to identify event arguments beyond the sentence level.
no code implementations • 16 Dec 2024 • Tianyi Yin, Jingwei Wang, Yunlong Ma, Han Wang, Chenze Wang, Yukai Zhao, Min Liu, Weiming Shen, Yufeng Chen
Encoding time series into tokens and using language models for processing has been shown to substantially augment the models' ability to generalize to unseen tasks.
no code implementations • 19 Sep 2024 • Chen Liang, Zhifan Feng, Zihe Liu, Wenbin Jiang, Jinan Xu, Yufeng Chen, Yong Wang
Chain-of-thought prompting significantly boosts the reasoning ability of large language models but still faces three issues: hallucination problem, restricted interpretability, and uncontrollable generation.
1 code implementation • 23 Jul 2024 • Yijie Chen, Yijin Liu, Fandong Meng, Jinan Xu, Yufeng Chen, Jie zhou
This study presents a benchmark AmbGIMT (Gender-Inclusive Machine Translation with Ambiguous attitude words), which assesses gender bias beyond binary gender.
1 code implementation • 25 Jun 2024 • Songming Zhang, Xue Zhang, Zengkui Sun, Yufeng Chen, Jinan Xu
Furthermore, this discrepancy also hinders the KD process between models with different vocabularies, which is common for current LLMs.
1 code implementation • 24 Jun 2024 • Xue Zhang, Yunlong Liang, Fandong Meng, Songming Zhang, Yufeng Chen, Jinan Xu, Jie zhou
To address this issue, we first investigate how LLMs process multilingual factual knowledge and discover that the same factual knowledge in different languages generally activates a shared set of neurons, which we call language-agnostic factual neurons (LAFNs).
1 code implementation • 5 Jun 2024 • Zengkui Sun, Yijin Liu, Jiaan Wang, Fandong Meng, Jinan Xu, Yufeng Chen, Jie zhou
Consequently, on the reasoning questions, we discover that existing methods struggle to utilize the edited knowledge to reason the new answer, and tend to retain outdated responses, which are generated by the original models utilizing original knowledge.
1 code implementation • 5 Jun 2024 • Zengkui Sun, Yijin Liu, Fandong Meng, Jinan Xu, Yufeng Chen, Jie zhou
Multilingual neural machine translation models generally distinguish translation directions by the language tag (LT) in front of the source or target sentences.
1 code implementation • 11 Apr 2024 • Yijie Chen, Yijin Liu, Fandong Meng, Yufeng Chen, Jinan Xu, Jie zhou
In this paper, we suggest that code comments are the natural logic pivot between natural language and code language and propose using comments to boost the code generation ability of code LLMs.
no code implementations • 9 Jan 2024 • Xue Zhang, Xiangyu Shi, Xinyue Lou, Rui Qi, Yufeng Chen, Jinan Xu, Wenjuan Han
Large language models (LLMs) and multimodal large language models (MLLMs) have shown excellent general capabilities, even exhibiting adaptability in many professional domains such as law, economics, transportation, and medicine.
no code implementations • 12 Dec 2023 • Xiangyu Shi, Yunlong Liang, Jinan Xu, Yufeng Chen
The results show that our method succeeds in reducing redundant retrieval operations and significantly reduces the overhead of kNN retrievals by up to 53% at the expense of a slight decline in translation quality.
no code implementations • 15 Nov 2023 • Yufeng Chen
It represents a significant step forward in machine learning, leveraging the inherent capabilities of GPT-4 to provide translations that are not only accurate but also contextually rich and linguistically sophisticated.
no code implementations • 5 Nov 2023 • Jiaxin Shen, Yanyao Liu, ZiMing Wang, Ziyuan Jiao, Yufeng Chen, Wenjuan Han
To facilitate the advancement of research in healthcare robots without human intervention or commands, we introduce the Autonomous Helping Challenge, along with a crowd-sourcing large-scale dataset.
1 code implementation • 20 Oct 2023 • Xue Zhang, Songming Zhang, Yunlong Liang, Yufeng Chen, Jian Liu, Wenjuan Han, Jinan Xu
Furthermore, for situations requiring multiple paraphrases for each source sentence, we design a Diverse Templates Search (DTS) algorithm, which can enhance the diversity between paraphrases without sacrificing quality.
1 code implementation • 24 Aug 2023 • Yijie Chen, Yijin Liu, Fandong Meng, Yufeng Chen, Jinan Xu, Jie zhou
The experimental results demonstrate significant improvements in translation performance with SWIE based on BLOOMZ-3b, particularly in zero-shot and long text translations due to reduced instruction forgetting risk.
1 code implementation • 3 Jul 2023 • Xiang Wei, Yufeng Chen, Ning Cheng, Xingyu Cui, Jinan Xu, Wenjuan Han
In order to construct or extend entity-centric and event-centric knowledge graphs (KG and EKG), the information extraction (IE) annotation toolkit is essential.
1 code implementation • 22 May 2023 • Yunlong Liang, Fandong Meng, Jiaan Wang, Jinan Xu, Yufeng Chen, Jie zhou
Further, we propose a dual knowledge distillation and target-oriented vision modeling framework for the M$^3$S task.
1 code implementation • 14 May 2023 • Songming Zhang, Yunlong Liang, Shuaibo Wang, Wenjuan Han, Jian Liu, Jinan Xu, Yufeng Chen
In this work, we first unravel this mystery from an empirical perspective and show that the knowledge comes from the top-1 predictions of teachers, which also helps us build a potential connection between word- and sequence-level KD.
no code implementations • 4 May 2023 • Yunlong Liang, Fandong Meng, Jinan Xu, Jiaan Wang, Yufeng Chen, Jie zhou
Specifically, we propose a ``versatile'' model, i. e., the Unified Model Learning for NMT (UMLNMT) that works with data from different tasks, and can translate well in multiple settings simultaneously, and theoretically it can be as many as possible.
no code implementations • 14 Apr 2023 • Yufeng Chen, Hongfei Dai, Wenlin Li, Fangmin Wang, Bo wang, Lijun Wang
It measures the clock difference between two locations without involving a data layer, which can reduce the complexity of the system.
1 code implementation • 20 Feb 2023 • Xiang Wei, Xingyu Cui, Ning Cheng, Xiaobin Wang, Xin Zhang, Shen Huang, Pengjun Xie, Jinan Xu, Yufeng Chen, Meishan Zhang, Yong Jiang, Wenjuan Han
Zero-shot information extraction (IE) aims to build IE systems from the unannotated text.
1 code implementation • 15 Dec 2022 • Yunlong Liang, Fandong Meng, Jinan Xu, Jiaan Wang, Yufeng Chen, Jie zhou
However, less attention has been paid to the visual features from the perspective of the summary, which may limit the model performance, especially in the low- and zero-resource scenarios.
no code implementations • 28 Nov 2022 • Yunlong Liang, Fandong Meng, Jinan Xu, Yufeng Chen, Jie zhou
Our systems achieve 0. 810 and 0. 946 COMET scores.
no code implementations • 12 Oct 2022 • Hongxiao Zhang, Siyu Lai, Songming Zhang, Hui Huang, Yufeng Chen, Jinan Xu, Jian Liu
This paper introduces the system used in our submission to the WMT'22 Translation Suggestion shared task.
1 code implementation • 9 Oct 2022 • Siyu Lai, Zhen Yang, Fandong Meng, Yufeng Chen, Jinan Xu, Jie zhou
Word alignment which aims to extract lexicon translation equivalents between source and target sentences, serves as a fundamental tool for natural language processing.
no code implementations • 20 Sep 2022 • Andrea Tagliabue, Yi-Hsuan Hsiao, Urban Fasel, J. Nathan Kutz, Steven L. Brunton, Yufeng Chen, Jonathan P. How
Accurate and agile trajectory tracking in sub-gram Micro Aerial Vehicles (MAVs) is challenging, as the small scale of the robot induces large model uncertainties, demanding robust feedback controllers, while the fast dynamics and computational constraints prevent the deployment of computationally expensive strategies.
1 code implementation • ACL 2022 • Yunlong Liang, Fandong Meng, Jinan Xu, Yufeng Chen, Jie zhou
Neural Chat Translation (NCT) aims to translate conversational text into different languages.
1 code implementation • NAACL 2022 • Siyu Lai, Zhen Yang, Fandong Meng, Xue Zhang, Yufeng Chen, Jinan Xu, Jie zhou
Generating adversarial examples for Neural Machine Translation (NMT) with single Round-Trip Translation (RTT) has achieved promising results by releasing the meaning-preserving restriction.
1 code implementation • ACL 2022 • Yunlong Liang, Fandong Meng, Chulun Zhou, Jinan Xu, Yufeng Chen, Jinsong Su, Jie zhou
The goal of the cross-lingual summarization (CLS) is to convert a document in one language (e. g., English) to a summary in another one (e. g., Chinese).
1 code implementation • ACL 2022 • Songming Zhang, Yijin Liu, Fandong Meng, Yufeng Chen, Jinan Xu, Jian Liu, Jie zhou
Token-level adaptive training approaches can alleviate the token imbalance problem and thus improve neural machine translation, through re-weighting the losses of different target tokens based on specific statistical metrics (e. g., token frequency or mutual information).
1 code implementation • ACL 2022 • Yunlong Liang, Fandong Meng, Jinan Xu, Yufeng Chen, Jie zhou
In this work, we introduce a new task named Multimodal Chat Translation (MCT), aiming to generate more accurate translations with the help of the associated dialogue history and visual context.
1 code implementation • EMNLP 2021 • Yunlong Liang, Chulun Zhou, Fandong Meng, Jinan Xu, Yufeng Chen, Jinsong Su, Jie zhou
Neural Chat Translation (NCT) aims to translate conversational text between speakers of different languages.
1 code implementation • EMNLP 2021 • Yijin Liu, Fandong Meng, Yufeng Chen, Jinan Xu, Jie zhou
Its core motivation is to simulate the inference scene during training by replacing ground-truth tokens with predicted tokens, thus bridging the gap between training and inference.
1 code implementation • ACL 2021 • Yunlong Liang, Fandong Meng, Yufeng Chen, Jinan Xu, Jie zhou
Despite the impressive performance of sentence-level and context-aware Neural Machine Translation (NMT), there still remain challenges to translate bilingual conversational text due to its inherent characteristics such as role preference, dialogue coherence, and translation consistency.
1 code implementation • Findings (ACL) 2021 • Yijin Liu, Fandong Meng, Yufeng Chen, Jinan Xu, Jie zhou
In this way, the model is exactly exposed to predicted tokens for high-confidence positions and still ground-truth tokens for low-confidence positions.
1 code implementation • Findings (ACL) 2021 • Ying Zhang, Fandong Meng, Yufeng Chen, Jinan Xu, Jie zhou
In this paper, we tackle the problem by transferring knowledge from three aspects, i. e., domain, language and task, and strengthening connections among them.
1 code implementation • 9 Dec 2020 • Yunlong Liang, Fandong Meng, Ying Zhang, Jinan Xu, Yufeng Chen, Jie zhou
Firstly, we design a Heterogeneous Graph-Based Encoder to represent the conversation content (i. e., the dialogue history, its emotion flow, facial expressions, audio, and speakers' personalities) with a heterogeneous graph neural network, and then predict suitable emotions for feedback.
no code implementations • Asian Chapter of the Association for Computational Linguistics 2020 • Wenbin Jiang, Mengfei Guo, Yufeng Chen, Ying Li, Jinan Xu, Yajuan Lyu, Yong Zhu
This paper describes a novel multi-view classification model for knowledge graph completion, where multiple classification views are performed based on both content and context information for candidate triple evaluation.
no code implementations • COLING 2020 • Mingtong Liu, Erguang Yang, Deyi Xiong, Yujie Zhang, Yao Meng, Changjian Hu, Jinan Xu, Yufeng Chen
We propose a learning-exploring method to generate sentences as learning objectives from the learned data distribution, and employ reinforcement learning to combine these new learning objectives for model training.
no code implementations • 12 Aug 2020 • Yunlong Liang, Fandong Meng, Jinchao Zhang, Yufeng Chen, Jinan Xu, Jie zhou
For multiple aspects scenario of aspect-based sentiment analysis (ABSA), existing approaches typically ignore inter-aspect relations or rely on temporal dependencies to process aspect-aware representations of all aspects in a sentence.
Aspect-Based Sentiment Analysis Aspect-Based Sentiment Analysis (ABSA) +1
no code implementations • 27 Apr 2020 • Yijin Liu, Fandong Meng, Jie zhou, Yufeng Chen, Jinan Xu
Depth-adaptive neural networks can dynamically adjust depths according to the hardness of input words, and thus improve efficiency.
2 code implementations • Findings (EMNLP) 2021 • Yunlong Liang, Fandong Meng, Jinchao Zhang, Yufeng Chen, Jinan Xu, Jie zhou
Aspect-based sentiment analysis (ABSA) mainly involves three subtasks: aspect term extraction, opinion term extraction, and aspect-level sentiment classification, which are typically handled in a separate or joint manner.
Aspect-Based Sentiment Analysis Aspect-Based Sentiment Analysis (ABSA) +3
3 code implementations • 4 Apr 2020 • Yunlong Liang, Fandong Meng, Jinchao Zhang, Jinan Xu, Yufeng Chen, Jie zhou
The aspect-based sentiment analysis (ABSA) task remains to be a long-standing challenge, which aims to extract the aspect term and then identify its sentiment orientation. In previous approaches, the explicit syntactic structure of a sentence, which reflects the syntax properties of natural language and hence is intuitively crucial for aspect term extraction and sentiment recognition, is typically neglected or insufficiently modeled.
Aspect-Based Sentiment Analysis Aspect-Based Sentiment Analysis (ABSA) +3
1 code implementation • 29 Feb 2020 • Yijin Liu, Fandong Meng, Yufeng Chen, Jinan Xu, Jie zhou
The Sentence-State LSTM (S-LSTM) is a powerful and high efficient graph recurrent network, which views words as nodes and performs layer-wise recurrent steps between them simultaneously.
no code implementations • IJCNLP 2019 • Mingtong Liu, Yu-Jie Zhang, Jinan Xu, Yufeng Chen
Unlike existing models, each attention layer of OSOA-DFN is oriented to the original semantic representation of another sentence, which captures the relevant information from a fixed matching target.
2 code implementations • IJCNLP 2019 • Yijin Liu, Fandong Meng, Jinchao Zhang, Jie zhou, Yufeng Chen, Jinan Xu
Spoken Language Understanding (SLU) mainly involves two tasks, intent detection and slot filling, which are generally modeled jointly in existing works.
Ranked #1 on Slot Filling on CAIS
1 code implementation • IJCNLP 2019 • Yunlong Liang, Fandong Meng, Jinchao Zhang, Jinan Xu, Yufeng Chen, Jie zhou
Aspect based sentiment analysis (ABSA) aims to identify the sentiment polarity towards the given aspect in a sentence, while previous models typically exploit an aspect-independent (weakly associative) encoder for sentence representation generation.
Aspect-Based Sentiment Analysis Aspect-Based Sentiment Analysis (ABSA) +1
1 code implementation • ACL 2019 • Yijin Liu, Fandong Meng, Jinchao Zhang, Jinan Xu, Yufeng Chen, Jie zhou
Current state-of-the-art systems for sequence labeling are typically based on the family of Recurrent Neural Networks (RNNs).
Ranked #20 on Named Entity Recognition (NER) on CoNLL 2003 (English) (using extra training data)
no code implementations • WS 2016 • Shaotong Li, Jinan Xu, Yufeng Chen, Yu-Jie Zhang
This paper presents our machine translation system that developed for the WAT2016 evalua-tion tasks of ja-en, ja-zh, en-ja, zh-ja, JPCja-en, JPCja-zh, JPCen-ja, JPCzh-ja.