Search Results for author: Shengjie Qiu

Found 6 papers, 6 papers with code

Spurious Forgetting in Continual Learning of Language Models

1 code implementation23 Jan 2025 Junhao Zheng, Xidi Cai, Shengjie Qiu, Qianli Ma

Recent advancements in large language models (LLMs) reveal a perplexing phenomenon in continual learning: despite extensive training, models experience significant performance declines, raising questions about task alignment and underlying knowledge retention.

Continual Learning

Towards Lifelong Learning of Large Language Models: A Survey

1 code implementation10 Jun 2024 Junhao Zheng, Shengjie Qiu, Chengming Shi, Qianli Ma

This survey delves into the sophisticated landscape of lifelong learning, categorizing strategies into two primary groups: Internal Knowledge and External Knowledge.

Continual Pretraining Incremental Learning +2

Incremental Sequence Labeling: A Tale of Two Shifts

2 code implementations16 Feb 2024 Shengjie Qiu, Junhao Zheng, Zhen Liu, Yicheng Luo, Qianli Ma

As for the E2O problem, we use knowledge distillation to maintain the model's discriminative ability for old entities.

Incremental Learning Knowledge Distillation

Can LLMs Learn New Concepts Incrementally without Forgetting?

2 code implementations13 Feb 2024 Junhao Zheng, Shengjie Qiu, Qianli Ma

The concepts in Concept-1K are discrete, interpretable units of knowledge that allow for fine-grained analysis of learning and forgetting processes.

In-Context Learning Incremental Learning

Learn or Recall? Revisiting Incremental Learning with Pre-trained Language Models

2 code implementations13 Dec 2023 Junhao Zheng, Shengjie Qiu, Qianli Ma

Most assume that catastrophic forgetting is the biggest obstacle to achieving superior IL performance and propose various techniques to overcome this issue.

Class Incremental Learning Incremental Learning +7

Preserving Commonsense Knowledge from Pre-trained Language Models via Causal Inference

2 code implementations19 Jun 2023 Junhao Zheng, Qianli Ma, Shengjie Qiu, Yue Wu, Peitian Ma, Junlong Liu, Huawen Feng, Xichen Shang, Haibin Chen

Intriguingly, the unified objective can be seen as the sum of the vanilla fine-tuning objective, which learns new knowledge from target data, and the causal objective, which preserves old knowledge from PLMs.

Attribute Causal Inference

Cannot find the paper you are looking for? You can Submit a new open access paper.