Search Results for author: Zixuan Ke

Found 20 papers, 13 papers with code

Bridging the Preference Gap between Retrievers and LLMs

no code implementations13 Jan 2024 Zixuan Ke, Weize Kong, Cheng Li, Mingyang Zhang, Qiaozhu Mei, Michael Bendersky

Large Language Models (LLMs) have demonstrated superior results across a wide range of tasks, and Retrieval-augmented Generation (RAG) is an effective way to enhance the performance by locating relevant information and placing it into the context window of the LLM.

Question Answering Retrieval

Parameter-Level Soft-Masking for Continual Learning

1 code implementation26 Jun 2023 Tatsuya Konishi, Mori Kurokawa, Chihiro Ono, Zixuan Ke, Gyuhak Kim, Bing Liu

Although several techniques have achieved learning with no CF, they attain it by letting each task monopolize a sub-network in a shared network, which seriously limits knowledge transfer (KT) and causes over-consumption of the network capacity, i. e., as more tasks are learned, the performance deteriorates.

Continual Learning Incremental Learning +1

Open-World Continual Learning: Unifying Novelty Detection and Continual Learning

no code implementations20 Apr 2023 Gyuhak Kim, Changnan Xiao, Tatsuya Konishi, Zixuan Ke, Bing Liu

The key theoretical result is that regardless of whether WP and OOD detection (or TP) are defined explicitly or implicitly by a CIL algorithm, good WP and good OOD detection are necessary and sufficient conditions for good CIL, which unifies novelty or OOD detection and continual learning (CIL, in particular).

Class Incremental Learning Incremental Learning +2

Adapting a Language Model While Preserving its General Knowledge

2 code implementations21 Jan 2023 Zixuan Ke, Yijia Shao, Haowei Lin, Hu Xu, Lei Shu, Bing Liu

This paper shows that the existing methods are suboptimal and proposes a novel method to perform a more informed adaptation of the knowledge in the LM by (1) soft-masking the attention heads based on their importance to best preserve the general knowledge in the LM and (2) contrasting the representations of the general and the full (both general and domain knowledge) to learn an integrated representation with both general and domain-specific knowledge.

Continual Learning General Knowledge +1

Continual Learning of Natural Language Processing Tasks: A Survey

1 code implementation23 Nov 2022 Zixuan Ke, Bing Liu

Continual learning (CL) is a learning paradigm that emulates the human capability of learning and accumulating knowledge continually without forgetting the previously learned knowledge and also transferring the learned knowledge to help learn new tasks better.

Continual Learning Transfer Learning

Continual Training of Language Models for Few-Shot Learning

3 code implementations11 Oct 2022 Zixuan Ke, Haowei Lin, Yijia Shao, Hu Xu, Lei Shu, Bing Liu

Recent work on applying large language models (LMs) achieves impressive performance in many NLP applications.

Continual Learning Continual Pretraining +2

A Multi-Head Model for Continual Learning via Out-of-Distribution Replay

3 code implementations20 Aug 2022 Gyuhak Kim, Zixuan Ke, Bing Liu

Instead of using the saved samples in memory to update the network for previous tasks/classes in the existing approach, MORE leverages the saved samples to build a task specific classifier (adding a new classification head) without updating the network learned for previous tasks/classes.

Class Incremental Learning Incremental Learning +1

Domain-Aware Contrastive Knowledge Transfer for Multi-domain Imbalanced Data

no code implementations WASSA (ACL) 2022 Zixuan Ke, Mohammad Kachuee, Sungjin Lee

In many real-world machine learning applications, samples belong to a set of domains e. g., for product reviews each review belongs to a product category.

Transfer Learning

Continual Learning of a Mixed Sequence of Similar and Dissimilar Tasks

2 code implementations NeurIPS 2020 Zixuan Ke, Bing Liu, Xingchang Huang

To the best of our knowledge, no technique has been proposed to learn a sequence of mixed similar and dissimilar tasks that can deal with forgetting and also transfer knowledge forward and backward.

Continual Learning

Continual Learning with Knowledge Transfer for Sentiment Classification

2 code implementations18 Dec 2021 Zixuan Ke, Bing Liu, Hao Wang, Lei Shu

In this setting, the CL system learns a sequence of SC tasks incrementally in a neural network, where each task builds a classifier to classify the sentiment of reviews of a particular product category or domain.

Classification Continual Learning +4

Achieving Forgetting Prevention and Knowledge Transfer in Continual Learning

1 code implementation NeurIPS 2021 Zixuan Ke, Bing Liu, Nianzu Ma, Hu Xu, Lei Shu

Although several papers have tried to deal with both CF and KT, our experiments show that they suffer from serious CF when the tasks do not have much shared knowledge.

Continual Learning Language Modelling +2

CLASSIC: Continual and Contrastive Learning of Aspect Sentiment Classification Tasks

1 code implementation EMNLP 2021 Zixuan Ke, Bing Liu, Hu Xu, Lei Shu

The key novelty is a contrastive continual learning method that enables both knowledge transfer across tasks and knowledge distillation from old tasks to the new task, which eliminates the need for task ids in testing.

Classification Continual Learning +6

Continual Learning Using Pseudo-Replay via Latent Space Sampling

no code implementations29 Sep 2021 Gyuhak Kim, Sepideh Esmaeilpour, Zixuan Ke, Tatsuya Konishi, Bing Liu

PLS is not only simple and efficient but also does not invade data privacy due to the fact that it works in the latent feature space.

Class Incremental Learning Incremental Learning

Partially Relaxed Masks for Lightweight Knowledge Transfer without Forgetting in Continual Learning

no code implementations29 Sep 2021 Tatsuya Konishi, Mori Kurokawa, Roberto Legaspi, Chihiro Ono, Zixuan Ke, Gyuhak Kim, Bing Liu

The goal of this work is to endow such systems with the additional ability to transfer knowledge among tasks when the tasks are similar and have shared knowledge to achieve higher accuracy.

Continual Learning Incremental Learning +1

Give Me More Feedback II: Annotating Thesis Strength and Related Attributes in Student Essays

no code implementations ACL 2019 Zixuan Ke, Hrishikesh Inamdar, Hui Lin, Vincent Ng

While the vast majority of existing work on automated essay scoring has focused on holistic scoring, researchers have recently begun work on scoring specific dimensions of essay quality.

Automated Essay Scoring

Give Me More Feedback: Annotating Argument Persuasiveness and Related Attributes in Student Essays

no code implementations ACL 2018 Winston Carlile, Nishant Gurrapadi, Zixuan Ke, Vincent Ng

While argument persuasiveness is one of the most important dimensions of argumentative essay quality, it is relatively little studied in automated essay scoring research.

Automated Essay Scoring Persuasiveness

Cannot find the paper you are looking for? You can Submit a new open access paper.