Search Results for author: Jia-Chen Gu

Found 37 papers, 24 papers with code

Conversation- and Tree-Structure Losses for Dialogue Disentanglement

no code implementations dialdoc (ACL) 2022 Tianda Li, Jia-Chen Gu, Zhen-Hua Ling, Quan Liu

When multiple conversations occur simultaneously, a listener must decide which conversation each utterance is part of in order to interpret and respond to it appropriately.

Disentanglement

Can Editing LLMs Inject Harm?

1 code implementation29 Jul 2024 Canyu Chen, Baixiang Huang, Zekun Li, Zhaorun Chen, Shiyang Lai, Xiongxiao Xu, Jia-Chen Gu, Jindong Gu, Huaxiu Yao, Chaowei Xiao, Xifeng Yan, William Yang Wang, Philip Torr, Dawn Song, Kai Shu

Then, we find that editing attacks can inject both types of misinformation into LLMs, and the effectiveness is particularly high for commonsense misinformation injection.

Fairness General Knowledge +4

Synchronous Faithfulness Monitoring for Trustworthy Retrieval-Augmented Generation

no code implementations19 Jun 2024 Di wu, Jia-Chen Gu, Fan Yin, Nanyun Peng, Kai-Wei Chang

Retrieval-augmented language models (RALMs) have shown strong performance and wide applicability in knowledge-intensive tasks.

Retrieval Uncertainty Quantification

Perturbation-Restrained Sequential Model Editing

no code implementations27 May 2024 Jun-Yu Ma, Hong Wang, Hao-Xiang Xu, Zhen-Hua Ling, Jia-Chen Gu

In this paper, we first theoretically analyze that the factor affecting the general abilities in sequential model editing lies in the condition number of the edited matrix.

Continual Learning Model Editing

Multiscale Matching Driven by Cross-Modal Similarity Consistency for Audio-Text Retrieval

no code implementations15 Mar 2024 Qian Wang, Jia-Chen Gu, Zhen-Hua Ling

Audio-text retrieval (ATR), which retrieves a relevant caption given an audio clip (A2T) and vice versa (T2A), has recently attracted much research attention.

AudioCaps Contrastive Learning +1

Neighboring Perturbations of Knowledge Editing on Large Language Models

1 code implementation31 Jan 2024 Jun-Yu Ma, Zhen-Hua Ling, Ningyu Zhang, Jia-Chen Gu

A metric of additivity is introduced and a benchmark dubbed as Perturbation Evaluation of Appending Knowledge (PEAK) is constructed to evaluate the degree of perturbation to neighboring knowledge when appending new knowledge.

knowledge editing

Corrective Retrieval Augmented Generation

2 code implementations29 Jan 2024 Shi-Qi Yan, Jia-Chen Gu, Yun Zhu, Zhen-Hua Ling

Experiments on four datasets covering short- and long-form generation tasks show that CRAG can significantly improve the performance of RAG-based approaches.

RAG Retrieval

Leveraging Large Language Models for NLG Evaluation: Advances and Challenges

1 code implementation13 Jan 2024 Zhen Li, Xiaohan Xu, Tao Shen, Can Xu, Jia-Chen Gu, Yuxuan Lai, Chongyang Tao, Shuai Ma

In the rapidly evolving domain of Natural Language Generation (NLG) evaluation, introducing Large Language Models (LLMs) has opened new avenues for assessing generated content quality, e. g., coherence, creativity, and context relevance.

nlg evaluation Specificity +1

Model Editing Harms General Abilities of Large Language Models: Regularization to the Rescue

1 code implementation9 Jan 2024 Jia-Chen Gu, Hao-Xiang Xu, Jun-Yu Ma, Pan Lu, Zhen-Hua Ling, Kai-Wei Chang, Nanyun Peng

While current model editing methods can effectively modify a model's behavior within a specific area of interest, they often overlook the potential unintended side effects on the general abilities of LLMs such as reasoning, natural language inference, and question answering.

Model Editing Natural Language Inference +1

Is ChatGPT a Good Multi-Party Conversation Solver?

1 code implementation25 Oct 2023 Chao-Hong Tan, Jia-Chen Gu, Zhen-Hua Ling

Large Language Models (LLMs) have emerged as influential instruments within the realm of natural language processing; nevertheless, their capacity to handle multi-party conversations (MPCs) -- a scenario marked by the presence of multiple interlocutors involved in intricate information exchanges -- remains uncharted.

Zero-Shot Learning

Untying the Reversal Curse via Bidirectional Language Model Editing

1 code implementation16 Oct 2023 Jun-Yu Ma, Jia-Chen Gu, Zhen-Hua Ling, Quan Liu, Cong Liu

A new evaluation metric of reversibility is introduced, and a benchmark dubbed as Bidirectional Assessment for Knowledge Editing (BAKE) is constructed to evaluate the reversibility of edited models in recalling knowledge in the reverse direction of editing.

knowledge editing Language Modelling +1

MADNet: Maximizing Addressee Deduction Expectation for Multi-Party Conversation Generation

1 code implementation22 May 2023 Jia-Chen Gu, Chao-Hong Tan, Caiyuan Chu, Zhen-Hua Ling, Chongyang Tao, Quan Liu, Cong Liu

Given an MPC with a few addressee labels missing, existing methods fail to build a consecutively connected conversation graph, but only a few separate conversation fragments instead.

SHINE: Syntax-augmented Hierarchical Interactive Encoder for Zero-shot Cross-lingual Information Extraction

no code implementations21 May 2023 Jun-Yu Ma, Jia-Chen Gu, Zhen-Hua Ling, Quan Liu, Cong Liu, Guoping Hu

The proposed encoder is capable of interactively capturing complementary information between features and contextual information, to derive language-agnostic representations for various IE tasks.

DiffuSIA: A Spiral Interaction Architecture for Encoder-Decoder Text Diffusion

no code implementations19 May 2023 Chao-Hong Tan, Jia-Chen Gu, Zhen-Hua Ling

In fact, the encoder-decoder architecture is naturally more flexible for its detachable encoder and decoder modules, which is extensible to multilingual and multimodal generation tasks for conditions and target texts.

Conditional Text Generation Decoder +5

GIFT: Graph-Induced Fine-Tuning for Multi-Party Conversation Understanding

1 code implementation16 May 2023 Jia-Chen Gu, Zhen-Hua Ling, Quan Liu, Cong Liu, Guoping Hu

Addressing the issues of who saying what to whom in multi-party conversations (MPCs) has recently attracted a lot of research attention.

Speaker Identification

Multi-Stage Coarse-to-Fine Contrastive Learning for Conversation Intent Induction

no code implementations9 Mar 2023 Caiyuan Chu, Ya Li, Yifan Liu, Jia-Chen Gu, Quan Liu, Yongxin Ge, Guoping Hu

The key to automatic intention induction is that, for any given set of new data, the sentence representation obtained by the model can be well distinguished from different labels.

Clustering Contrastive Learning +3

WIDER & CLOSER: Mixture of Short-channel Distillers for Zero-shot Cross-lingual Named Entity Recognition

1 code implementation7 Dec 2022 Jun-Yu Ma, Beiduo Chen, Jia-Chen Gu, Zhen-Hua Ling, Wu Guo, Quan Liu, Zhigang Chen, Cong Liu

In this study, a mixture of short-channel distillers (MSD) method is proposed to fully interact the rich hierarchical information in the teacher model and to transfer knowledge to the student model sufficiently and efficiently.

Cross-Lingual NER Domain Adaptation +3

HeterMPC: A Heterogeneous Graph Neural Network for Response Generation in Multi-Party Conversations

1 code implementation ACL 2022 Jia-Chen Gu, Chao-Hong Tan, Chongyang Tao, Zhen-Hua Ling, Huang Hu, Xiubo Geng, Daxin Jiang

To address these challenges, we present HeterMPC, a heterogeneous graph-based neural network for response generation in MPCs which models the semantics of utterances and interlocutors simultaneously with two types of nodes in a graph.

Graph Neural Network Response Generation

Detecting Speaker Personas from Conversational Texts

1 code implementation EMNLP 2021 Jia-Chen Gu, Zhen-Hua Ling, Yu Wu, Quan Liu, Zhigang Chen, Xiaodan Zhu

This is a many-to-many semantic matching task because both contexts and personas in SPD are composed of multiple sentences.

MPC-BERT: A Pre-Trained Language Model for Multi-Party Conversation Understanding

1 code implementation ACL 2021 Jia-Chen Gu, Chongyang Tao, Zhen-Hua Ling, Can Xu, Xiubo Geng, Daxin Jiang

Recently, various neural models for multi-party conversation (MPC) have achieved impressive improvements on a variety of tasks such as addressee recognition, speaker identification and response prediction.

Language Modelling Speaker Identification

Partner Matters! An Empirical Study on Fusing Personas for Personalized Response Selection in Retrieval-Based Chatbots

1 code implementation19 May 2021 Jia-Chen Gu, Hui Liu, Zhen-Hua Ling, Quan Liu, Zhigang Chen, Xiaodan Zhu

Empirical studies on the Persona-Chat dataset show that the partner personas neglected in previous studies can improve the accuracy of response selection in the IMN- and BERT-based models.

Retrieval

Learning to Retrieve Entity-Aware Knowledge and Generate Responses with Copy Mechanism for Task-Oriented Dialogue Systems

1 code implementation22 Dec 2020 Chao-Hong Tan, Xiaoyu Yang, Zi'ou Zheng, Tianda Li, Yufei Feng, Jia-Chen Gu, Quan Liu, Dan Liu, Zhen-Hua Ling, Xiaodan Zhu

Task-oriented conversational modeling with unstructured knowledge access, as track 1 of the 9th Dialogue System Technology Challenges (DSTC 9), requests to build a system to generate response given dialogue history and knowledge access.

Response Generation Task-Oriented Dialogue Systems

Filtering before Iteratively Referring for Knowledge-Grounded Response Selection in Retrieval-Based Chatbots

1 code implementation Findings of the Association for Computational Linguistics 2020 Jia-Chen Gu, Zhen-Hua Ling, Quan Liu, Zhigang Chen, Xiaodan Zhu

The challenges of building knowledge-grounded retrieval-based chatbots lie in how to ground a conversation on its background knowledge and how to match response candidates with both context and knowledge simultaneously.

Retrieval

DialBERT: A Hierarchical Pre-Trained Model for Conversation Disentanglement

1 code implementation8 Apr 2020 Tianda Li, Jia-Chen Gu, Xiaodan Zhu, Quan Liu, Zhen-Hua Ling, Zhiming Su, Si Wei

Disentanglement is a problem in which multiple conversations occur in the same channel simultaneously, and the listener should decide which utterance is part of the conversation he will respond to.

Conversation Disentanglement Disentanglement

Dually Interactive Matching Network for Personalized Response Selection in Retrieval-Based Chatbots

1 code implementation IJCNLP 2019 Jia-Chen Gu, Zhen-Hua Ling, Xiaodan Zhu, Quan Liu

Compared with previous persona fusion approaches which enhance the representation of a context by calculating its similarity with a given persona, the DIM model adopts a dual matching architecture, which performs interactive matching between responses and contexts and between responses and personas respectively for ranking response candidates.

Retrieval

Promoting Diversity for End-to-End Conversation Response Generation

no code implementations27 Jan 2019 Yu-Ping Ruan, Zhen-Hua Ling, Quan Liu, Jia-Chen Gu, Xiaodan Zhu

At this stage, two different models are proposed, i. e., a variational generative (VariGen) model and a retrieval based (Retrieval) model.

Diversity Response Generation +1

Cannot find the paper you are looking for? You can Submit a new open access paper.