Search Results for author: Jinan Xu

Found 42 papers, 18 papers with code

基于多任务标签一致性机制的中文命名实体识别(Chinese Named Entity Recognition based on Multi-task Label Consistency Mechanism)

no code implementations CCL 2021 Shuning Lv, Jian Liu, Jinan Xu, Yufeng Chen, Yujie Zhang

“实体边界预测对中文命名实体识别至关重要。现有研究为改善边界识别效果提出的多任务学习方法仅考虑与分词任务结合, 缺少多任务标签训练数据, 无法学到任务的标签一致性关系。本文提出一种新的基于多任务标签一致性机制的中文命名实体识别方法:将分词和词性信息融入命名实体识别模型, 使三种任务联合训练;建立基于标签一致性机制的多任务学习模式, 来捕获标签一致性关系及学习多任务表示。全样本和小样本实验表明了方法的有效性。”

Chinese Named Entity Recognition named-entity-recognition

Syntactically Diverse Adversarial Network for Knowledge-Grounded Conversation Generation

no code implementations Findings (EMNLP) 2021 Fuwei Cui, Hui Di, Hongjie Ren, Kazushige Ouchi, Ze Liu, Jinan Xu

Generative conversation systems tend to produce meaningless and generic responses, which significantly reduce the user experience.

Informativeness

融合外部知识的开放域复述模板获取方法(An Open Domain Paraphrasing Template Acquisition Method Based on External Knowledge)

no code implementations CCL 2021 Bo Jin, Mingtong Liu, Yujie Zhang, Jinan Xu, Yufeng Chen

“如何挖掘语言资源中丰富的复述模板, 是复述研究中的一项重要任务。已有方法在人工给定种子实体对的基础上, 利用实体关系, 通过自举迭代方式, 从开放域获取复述模板, 规避对平行语料或可比语料的依赖, 但是该方法需人工给定实体对, 实体关系受限;在迭代过程中语义会发生偏移, 影响获取质量。针对这些问题, 我们考虑知识库中包含描述特定语义关系的实体对(即关系三元组), 提出融合外部知识的开放域复述模板自动获取方法。首先, 将关系三元组与开放域文本对齐, 获取关系对应文本, 并将文本中语义丰富部分泛化成变量槽, 获取关系模板;接着设计模板表示方法, 本文利用预训练语言模型, 在模板表示中融合变量槽语义;最后, 根据获得的模板表示, 设计自动聚类与筛选方法, 获取高精度的复述模板。在融合自动评测与人工评测的评价方法下, 实验结果表明, 本文提出的方法实现了在开放域数据上复述模板的自动泛化与获取, 能够获得质量高、语义一致的复述模板。”

Saliency as Evidence: Event Detection with Trigger Saliency Attribution

no code implementations ACL 2022 Jian Liu, Yufeng Chen, Jinan Xu

Event detection (ED) is a critical subtask of event extraction that seeks to identify event triggers of certain types in texts. Despite significant advances in ED, existing methods typically follow a “one model fits all types” approach, which sees no differences between event types and often results in a quite skewed performance. Finding the causes of skewed performance is crucial for the robustness of an ED model, but to date there has been little exploration of this problem. This research examines the issue in depth and presents a new concept termed trigger salience attribution, which can explicitly quantify the underlying patterns of events.

Event Detection Event Extraction

A Joint Model for Graph-based Chinese Dependency Parsing

no code implementations CCL 2020 Xingchen Li, Mingtong Liu, Yujie Zhang, Jinan Xu, Yufeng Chen

The experimental results on the Penn Chinese treebank (CTB5) show that our proposed joint model improved by 0. 38% on dependency parsing than the model of Yan et al. (2019).

Chinese Dependency Parsing Chinese Word Segmentation +3

Machine Reading Comprehension as Data Augmentation: A Case Study on Implicit Event Argument Extraction

no code implementations EMNLP 2021 Jian Liu, Yufeng Chen, Jinan Xu

Implicit event argument extraction (EAE) is a crucial document-level information extraction task that aims to identify event arguments beyond the sentence level.

Data Augmentation Event Argument Extraction +2

Generating Authentic Adversarial Examples beyond Meaning-preserving with Doubly Round-trip Translation

no code implementations19 Apr 2022 Siyu Lai, Zhen Yang, Fandong Meng, Xue Zhang, Yufeng Chen, Jinan Xu, Jie zhou

Generating adversarial examples for Neural Machine Translation (NMT) with single Round-Trip Translation (RTT) has achieved promising results by releasing the meaning-preserving restriction.

Machine Translation Translation

A Variational Hierarchical Model for Neural Cross-Lingual Summarization

1 code implementation ACL 2022 Yunlong Liang, Fandong Meng, Chulun Zhou, Jinan Xu, Yufeng Chen, Jinsong Su, Jie zhou

The goal of the cross-lingual summarization (CLS) is to convert a document in one language (e. g., English) to a summary in another one (e. g., Chinese).

Machine Translation Translation

Conditional Bilingual Mutual Information Based Adaptive Training for Neural Machine Translation

1 code implementation ACL 2022 Songming Zhang, Yijin Liu, Fandong Meng, Yufeng Chen, Jinan Xu, Jian Liu, Jie zhou

Token-level adaptive training approaches can alleviate the token imbalance problem and thus improve neural machine translation, through re-weighting the losses of different target tokens based on specific statistical metrics (e. g., token frequency or mutual information).

Language Modelling Machine Translation +1

MSCTD: A Multimodal Sentiment Chat Translation Dataset

1 code implementation ACL 2022 Yunlong Liang, Fandong Meng, Jinan Xu, Yufeng Chen, Jie zhou

In this work, we introduce a new task named Multimodal Chat Translation (MCT), aiming to generate more accurate translations with the help of the associated dialogue history and visual context.

Multimodal Machine Translation Sentiment Analysis +1

Scheduled Sampling Based on Decoding Steps for Neural Machine Translation

1 code implementation EMNLP 2021 Yijin Liu, Fandong Meng, Yufeng Chen, Jinan Xu, Jie zhou

Its core motivation is to simulate the inference scene during training by replacing ground-truth tokens with predicted tokens, thus bridging the gap between training and inference.

Machine Translation Text Summarization +1

WeChat Neural Machine Translation Systems for WMT21

no code implementations WMT (EMNLP) 2021 Xianfeng Zeng, Yijin Liu, Ernan Li, Qiu Ran, Fandong Meng, Peng Li, Jinan Xu, Jie zhou

This paper introduces WeChat AI's participation in WMT 2021 shared news translation task on English->Chinese, English->Japanese, Japanese->English and English->German.

Knowledge Distillation Machine Translation +3

Modeling Bilingual Conversational Characteristics for Neural Chat Translation

1 code implementation ACL 2021 Yunlong Liang, Fandong Meng, Yufeng Chen, Jinan Xu, Jie zhou

Despite the impressive performance of sentence-level and context-aware Neural Machine Translation (NMT), there still remain challenges to translate bilingual conversational text due to its inherent characteristics such as role preference, dialogue coherence, and translation consistency.

Machine Translation Translation

Confidence-Aware Scheduled Sampling for Neural Machine Translation

1 code implementation Findings (ACL) 2021 Yijin Liu, Fandong Meng, Yufeng Chen, Jinan Xu, Jie zhou

In this way, the model is exactly exposed to predicted tokens for high-confidence positions and still ground-truth tokens for low-confidence positions.

Machine Translation Translation

Target-Oriented Fine-tuning for Zero-Resource Named Entity Recognition

1 code implementation Findings (ACL) 2021 Ying Zhang, Fandong Meng, Yufeng Chen, Jinan Xu, Jie zhou

In this paper, we tackle the problem by transferring knowledge from three aspects, i. e., domain, language and task, and strengthening connections among them.

named-entity-recognition NER +1

Bilingual Mutual Information Based Adaptive Training for Neural Machine Translation

1 code implementation ACL 2021 Yangyifan Xu, Yijin Liu, Fandong Meng, Jiajun Zhang, Jinan Xu, Jie zhou

Recently, token-level adaptive training has achieved promising improvement in machine translation, where the cross-entropy loss function is adjusted by assigning different training weights to different tokens, in order to alleviate the token imbalance problem.

Machine Translation Translation

Emotional Conversation Generation with Heterogeneous Graph Neural Network

1 code implementation9 Dec 2020 Yunlong Liang, Fandong Meng, Ying Zhang, Jinan Xu, Yufeng Chen, Jie zhou

Firstly, we design a Heterogeneous Graph-Based Encoder to represent the conversation content (i. e., the dialogue history, its emotion flow, facial expressions, audio, and speakers' personalities) with a heterogeneous graph neural network, and then predict suitable emotions for feedback.

Multi-view Classification Model for Knowledge Graph Completion

no code implementations Asian Chapter of the Association for Computational Linguistics 2020 Wenbin Jiang, Mengfei Guo, Yufeng Chen, Ying Li, Jinan Xu, Yajuan Lyu, Yong Zhu

This paper describes a novel multi-view classification model for knowledge graph completion, where multiple classification views are performed based on both content and context information for candidate triple evaluation.

Classification Knowledge Graph Completion

A Learning-Exploring Method to Generate Diverse Paraphrases with Multi-Objective Deep Reinforcement Learning

no code implementations COLING 2020 Mingtong Liu, Erguang Yang, Deyi Xiong, Yujie Zhang, Yao Meng, Changjian Hu, Jinan Xu, Yufeng Chen

We propose a learning-exploring method to generate sentences as learning objectives from the learned data distribution, and employ reinforcement learning to combine these new learning objectives for model training.

Natural Language Processing Paraphrase Generation +1

Modeling Inter-Aspect Dependencies with a Non-temporal Mechanism for Aspect-Based Sentiment Analysis

no code implementations12 Aug 2020 Yunlong Liang, Fandong Meng, Jinchao Zhang, Yufeng Chen, Jinan Xu, Jie zhou

For multiple aspects scenario of aspect-based sentiment analysis (ABSA), existing approaches typically ignore inter-aspect relations or rely on temporal dependencies to process aspect-aware representations of all aspects in a sentence.

Aspect-Based Sentiment Analysis

Faster Depth-Adaptive Transformers

no code implementations27 Apr 2020 Yijin Liu, Fandong Meng, Jie zhou, Yufeng Chen, Jinan Xu

Depth-adaptive neural networks can dynamically adjust depths according to the hardness of input words, and thus improve efficiency.

Sentence Embeddings Text Classification

An Iterative Multi-Knowledge Transfer Network for Aspect-Based Sentiment Analysis

2 code implementations Findings (EMNLP) 2021 Yunlong Liang, Fandong Meng, Jinchao Zhang, Yufeng Chen, Jinan Xu, Jie zhou

Aspect-based sentiment analysis (ABSA) mainly involves three subtasks: aspect term extraction, opinion term extraction, and aspect-level sentiment classification, which are typically handled in a separate or joint manner.

Aspect-Based Sentiment Analysis Term Extraction +1

A Dependency Syntactic Knowledge Augmented Interactive Architecture for End-to-End Aspect-based Sentiment Analysis

3 code implementations4 Apr 2020 Yunlong Liang, Fandong Meng, Jinchao Zhang, Jinan Xu, Yufeng Chen, Jie zhou

The aspect-based sentiment analysis (ABSA) task remains to be a long-standing challenge, which aims to extract the aspect term and then identify its sentiment orientation. In previous approaches, the explicit syntactic structure of a sentence, which reflects the syntax properties of natural language and hence is intuitively crucial for aspect term extraction and sentiment recognition, is typically neglected or insufficiently modeled.

Aspect-Based Sentiment Analysis Multi-Task Learning +1

Depth-Adaptive Graph Recurrent Network for Text Classification

1 code implementation29 Feb 2020 Yijin Liu, Fandong Meng, Yufeng Chen, Jinan Xu, Jie zhou

The Sentence-State LSTM (S-LSTM) is a powerful and high efficient graph recurrent network, which views words as nodes and performs layer-wise recurrent steps between them simultaneously.

Classification General Classification +1

Original Semantics-Oriented Attention and Deep Fusion Network for Sentence Matching

no code implementations IJCNLP 2019 Mingtong Liu, Yu-Jie Zhang, Jinan Xu, Yufeng Chen

Unlike existing models, each attention layer of OSOA-DFN is oriented to the original semantic representation of another sentence, which captures the relevant information from a fixed matching target.

Natural Language Inference Paraphrase Identification

CM-Net: A Novel Collaborative Memory Network for Spoken Language Understanding

1 code implementation IJCNLP 2019 Yijin Liu, Fandong Meng, Jinchao Zhang, Jie zhou, Yufeng Chen, Jinan Xu

Spoken Language Understanding (SLU) mainly involves two tasks, intent detection and slot filling, which are generally modeled jointly in existing works.

Intent Detection Slot Filling +1

A Novel Aspect-Guided Deep Transition Model for Aspect Based Sentiment Analysis

1 code implementation IJCNLP 2019 Yunlong Liang, Fandong Meng, Jinchao Zhang, Jinan Xu, Yufeng Chen, Jie zhou

Aspect based sentiment analysis (ABSA) aims to identify the sentiment polarity towards the given aspect in a sentence, while previous models typically exploit an aspect-independent (weakly associative) encoder for sentence representation generation.

Aspect-Based Sentiment Analysis

GCDT: A Global Context Enhanced Deep Transition Architecture for Sequence Labeling

1 code implementation ACL 2019 Yijin Liu, Fandong Meng, Jinchao Zhang, Jinan Xu, Yufeng Chen, Jie zhou

Current state-of-the-art systems for sequence labeling are typically based on the family of Recurrent Neural Networks (RNNs).

Ranked #14 on Named Entity Recognition on CoNLL 2003 (English) (using extra training data)

Chunking NER +1

System Description of bjtu\_nlp Neural Machine Translation System

no code implementations WS 2016 Shaotong Li, Jinan Xu, Yufeng Chen, Yu-Jie Zhang

This paper presents our machine translation system that developed for the WAT2016 evalua-tion tasks of ja-en, ja-zh, en-ja, zh-ja, JPCja-en, JPCja-zh, JPCen-ja, JPCzh-ja.

Machine Translation Translation +1

Cannot find the paper you are looking for? You can Submit a new open access paper.