Search Results for author: Kun-Peng Ning

Found 10 papers, 4 papers with code

Sparse Orthogonal Parameters Tuning for Continual Learning

no code implementations5 Nov 2024 Kun-Peng Ning, Hai-Jian Ke, Yu-Yang Liu, Jia-Yu Yao, Yong-Hong Tian, Li Yuan

We hypothesize that the effectiveness of SoTU lies in the transformation of knowledge learned from multiple domains into the fusion of orthogonal delta parameters.

Continual Learning

Is Parameter Collision Hindering Continual Learning in LLMs?

no code implementations14 Oct 2024 Shuo Yang, Kun-Peng Ning, Yu-Yang Liu, Jia-Yu Yao, Yong-Hong Tian, Yi-Bing Song, Li Yuan

Large Language Models (LLMs) often suffer from catastrophic forgetting when learning multiple tasks sequentially, making continual learning (CL) essential for their dynamic deployment.

Continual Learning

Bidirectional Uncertainty-Based Active Learning for Open Set Annotation

1 code implementation23 Feb 2024 Chen-Chen Zong, Ye-Wen Wang, Kun-Peng Ning, Hai-Bo Ye, Sheng-Jun Huang

In this paper, we attempt to query examples that are both likely from known classes and highly informative, and propose a Bidirectional Uncertainty-based Active Learning (BUAL) framework.

Active Learning

PiCO: Peer Review in LLMs based on the Consistency Optimization

1 code implementation2 Feb 2024 Kun-Peng Ning, Shuo Yang, Yu-Yang Liu, Jia-Yu Yao, Zhen-Hui Liu, Yu Wang, Ming Pang, Li Yuan

Existing large language models (LLMs) evaluation methods typically focus on testing the performance on some closed-environment and domain-specific benchmarks with human annotations.

PICO

LLM Lies: Hallucinations are not Bugs, but Features as Adversarial Examples

1 code implementation2 Oct 2023 Jia-Yu Yao, Kun-Peng Ning, Zhen-Hui Liu, Mu-Nan Ning, Yu-Yang Liu, Li Yuan

This phenomenon forces us to revisit that \emph{hallucination may be another view of adversarial examples}, and it shares similar characteristics with conventional adversarial examples as a basic property of LLMs.

Hallucination

Towards Better Query Classification with Multi-Expert Knowledge Condensation in JD Ads Search

no code implementations2 Aug 2023 Kun-Peng Ning, Ming Pang, Zheng Fang, Xue Jiang, Xi-Wei Zhao, Chang-Ping Peng, Zhan-Gang Lin, Jing-He Hu, Jing-Ping Shao

To overcome this challenge, in this paper, we propose knowledge condensation (KC), a simple yet effective knowledge distillation framework to boost the classification performance of the online FastText model under strict low latency constraints.

Knowledge Distillation

Active Learning for Open-set Annotation

2 code implementations CVPR 2022 Kun-Peng Ning, Xun Zhao, Yu Li, Sheng-Jun Huang

To tackle this open-set annotation (OSA) problem, we propose a new active learning framework called LfOSA, which boosts the classification performance with an effective sampling strategy to precisely detect examples from known classes for annotation.

Active Learning

Improving Model Robustness by Adaptively Correcting Perturbation Levels with Active Queries

no code implementations27 Mar 2021 Kun-Peng Ning, Lue Tao, Songcan Chen, Sheng-Jun Huang

Recently, much research has been devoted to improving the model robustness by training with noise perturbations.

Active Learning

Co-Imitation Learning without Expert Demonstration

no code implementations27 Mar 2021 Kun-Peng Ning, Hu Xu, Kun Zhu, Sheng-Jun Huang

Imitation learning is a primary approach to improve the efficiency of reinforcement learning by exploiting the expert demonstrations.

Imitation Learning

Reinforcement Learning with Supervision from Noisy Demonstrations

no code implementations14 Jun 2020 Kun-Peng Ning, Sheng-Jun Huang

In this paper, we propose a novel framework to adaptively learn the policy by jointly interacting with the environment and exploiting the expert demonstrations.

reinforcement-learning Reinforcement Learning +1

Cannot find the paper you are looking for? You can Submit a new open access paper.