Search Results for author: Chengming Li

Found 17 papers, 6 papers with code

Amalgamating Knowledge from Two Teachers for Task-oriented Dialogue System with Adversarial Training

1 code implementation EMNLP 2020 Wanwei He, Min Yang, Rui Yan, Chengming Li, Ying Shen, Ruifeng Xu

Instead of adopting the classic student-teacher learning of forcing the output of a student network to exactly mimic the soft targets produced by the teacher networks, we introduce two discriminators as in generative adversarial network (GAN) to transfer knowledge from two teachers to the student.

Generative Adversarial Network Task-Oriented Dialogue Systems

CLHA: A Simple yet Effective Contrastive Learning Framework for Human Alignment

no code implementations25 Mar 2024 Feiteng Fang, Liang Zhu, Min Yang, Xi Feng, Jinchang Hou, Qixuan Zhao, Chengming Li, Xiping Hu, Ruifeng Xu

Reinforcement learning from human feedback (RLHF) is a crucial technique in aligning large language models (LLMs) with human preferences, ensuring these LLMs behave in beneficial and comprehensible ways to users.

Contrastive Learning reinforcement-learning

MoZIP: A Multilingual Benchmark to Evaluate Large Language Models in Intellectual Property

1 code implementation26 Feb 2024 Shiwen Ni, Minghuan Tan, Yuelin Bai, Fuqiang Niu, Min Yang, BoWen Zhang, Ruifeng Xu, Xiaojun Chen, Chengming Li, Xiping Hu, Ye Li, Jianping Fan

In this paper, we contribute a new benchmark, the first Multilingual-oriented quiZ on Intellectual Property (MoZIP), for the evaluation of LLMs in the IP domain.

Language Modelling Large Language Model +2

Layer-wise Regularized Dropout for Neural Language Models

no code implementations26 Feb 2024 Shiwen Ni, Min Yang, Ruifeng Xu, Chengming Li, Xiping Hu

To solve the inconsistency between training and inference caused by the randomness of dropout, some studies use consistency training to regularize dropout at the output layer.

Abstractive Text Summarization Machine Translation +1

History, Development, and Principles of Large Language Models-An Introductory Survey

no code implementations10 Feb 2024 Zhibo Chu, Shiwen Ni, Zichong Wang, Xi Feng, Chengming Li, Xiping Hu, Ruifeng Xu, Min Yang, Wenbin Zhang

Language models serve as a cornerstone in natural language processing (NLP), utilizing mathematical methods to generalize language laws and knowledge for prediction and generation.

Language Modelling

E-EVAL: A Comprehensive Chinese K-12 Education Evaluation Benchmark for Large Language Models

1 code implementation29 Jan 2024 Jinchang Hou, Chang Ao, Haihong Wu, Xiangtao Kong, Zhigang Zheng, Daijia Tang, Chengming Li, Xiping Hu, Ruifeng Xu, Shiwen Ni, Min Yang

The integration of LLMs and education is getting closer and closer, however, there is currently no benchmark for evaluating LLMs that focuses on the Chinese K-12 education domain.

Ethics Multiple-choice

Forgetting before Learning: Utilizing Parametric Arithmetic for Knowledge Updating in Large Language Models

no code implementations14 Nov 2023 Shiwen Ni, Dingwei Chen, Chengming Li, Xiping Hu, Ruifeng Xu, Min Yang

In this paper, we propose a new paradigm for fine-tuning called F-Learning (Forgetting before Learning), which employs parametric arithmetic to facilitate the forgetting of old knowledge and learning of new knowledge.

Expression Syntax Information Bottleneck for Math Word Problems

1 code implementation24 Oct 2023 Jing Xiong, Chengming Li, Min Yang, Xiping Hu, Bin Hu

To this end, we design an Expression Syntax Information Bottleneck method for MWP (called ESIB) based on variational information bottleneck, which extracts essential features of expression syntax tree while filtering latent-specific redundancy containing syntax-irrelevant features.

Math

DQ-LoRe: Dual Queries with Low Rank Approximation Re-ranking for In-Context Learning

1 code implementation4 Oct 2023 Jing Xiong, Zixuan Li, Chuanyang Zheng, Zhijiang Guo, Yichun Yin, Enze Xie, Zhicheng Yang, Qingxing Cao, Haiming Wang, Xiongwei Han, Jing Tang, Chengming Li, Xiaodan Liang

Dual Queries first query LLM to obtain LLM-generated knowledge such as CoT, then query the retriever to obtain the final exemplars via both question and the knowledge.

Dimensionality Reduction In-Context Learning +1

Self-Distillation with Meta Learning for Knowledge Graph Completion

1 code implementation Findings of the Association for Computational Linguistics: EMNLP 2022 2022 Yunshui Li, Junhao Liu, Chengming Li, Min Yang

In this paper, we propose a selfdistillation framework with meta learning(MetaSD) for knowledge graph completion with dynamic pruning, which aims to learn compressed graph embeddings and tackle the longtail samples.

Knowledge Graph Completion Meta-Learning +1

Self-consistent Reasoning For Solving Math Word Problems

no code implementations27 Oct 2022 Jing Xiong, Zhongwei Wan, Xiping Hu, Min Yang, Chengming Li

Specifically, we firstly obtain a sub-network by pruning a roberta2tree model, for the sake to use the gap on output distribution between the original roberta2tree model and the pruned sub-network to expose spurious correlative samples.

Math

Scene-adaptive Knowledge Distillation for Sequential Recommendation via Differentiable Architecture Search

no code implementations15 Jul 2021 Lei Chen, Fajie Yuan, Jiaxi Yang, Min Yang, Chengming Li

To realize such a goal, we propose AdaRec, a knowledge distillation (KD) framework which compresses knowledge of a teacher model into a student model adaptively according to its recommendation scene by using differentiable Neural Architecture Search (NAS).

Knowledge Distillation Neural Architecture Search +1

More than Encoder: Introducing Transformer Decoder to Upsample

no code implementations20 Jun 2021 Yijiang Li, Wentian Cai, Ying Gao, Chengming Li, Xiping Hu

The local and detailed feature from the shallower layer such as boundary and tissue texture is particularly more important in medical segmentation compared with natural image segmentation.

Image Segmentation Medical Image Segmentation +3

User-specific Adaptive Fine-tuning for Cross-domain Recommendations

no code implementations15 Jun 2021 Lei Chen, Fajie Yuan, Jiaxi Yang, Xiangnan He, Chengming Li, Min Yang

Fine-tuning works as an effective transfer learning technique for this objective, which adapts the parameters of a pre-trained model from the source domain to the target domain.

Recommendation Systems Transfer Learning

Interactive Key-Value Memory-augmented Attention for Image Paragraph Captioning

no code implementations COLING 2020 Chunpu Xu, Yu Li, Chengming Li, Xiang Ao, Min Yang, Jinwen Tian

In this paper, we propose an Interactive key-value Memory- augmented Attention model for image Paragraph captioning (IMAP) to keep track of the attention history (salient objects coverage information) along with the update-chain of the decoder state and therefore avoid generating repetitive or incomplete image descriptions.

Image Paragraph Captioning

Cannot find the paper you are looking for? You can Submit a new open access paper.