Search Results for author: Wenhao Liu

Found 29 papers, 14 papers with code

Few-Shot Intent Classification by Gauging Entailment Relationship Between Utterance and Semantic Label

no code implementations EMNLP (NLP4ConvAI) 2021 Jin Qu, Kazuma Hashimoto, Wenhao Liu, Caiming Xiong, Yingbo Zhou

Compared with DNNC, our proposed method is more efficient in both training and serving since it is based upon the entailment between query utterance and labels instead of all the training examples.

Classification intent-classification +2

Simple Data Augmentation with the Mask Token Improves Domain Adaptation for Dialog Act Tagging

no code implementations EMNLP 2020 Semih Yavuz, Kazuma Hashimoto, Wenhao Liu, Nitish Shirish Keskar, Richard Socher, Caiming Xiong

The concept of Dialogue Act (DA) is universal across different task-oriented dialogue domains - the act of {``}request{''} carries the same speaker intention whether it is for restaurant reservation or flight booking.

Data Augmentation Domain Generalization

Advancing Parameter Efficiency in Fine-tuning via Representation Editing

no code implementations23 Feb 2024 Muling Wu, Wenhao Liu, Xiaohua Wang, Tianlong Li, Changze Lv, Zixuan Ling, Jianhao Zhu, Cenyuan Zhang, Xiaoqing Zheng, Xuanjing Huang

Parameter Efficient Fine-Tuning (PEFT) has gained significant attention for its ability to achieve competitive results while updating only a small subset of trainable parameters.

Open the Pandora's Box of LLMs: Jailbreaking LLMs through Representation Engineering

no code implementations12 Jan 2024 Tianlong Li, Shihan Dou, Wenhao Liu, Muling Wu, Changze Lv, Xiaoqing Zheng, Xuanjing Huang

To overcome these limitations, we propose a novel jailbreaking approach, named Jailbreaking LLMs through Representation Engineering (JRE).

Prompt Engineering

Aligning Large Language Models with Human Preferences through Representation Engineering

no code implementations26 Dec 2023 Wenhao Liu, Xiaohua Wang, Muling Wu, Tianlong Li, Changze Lv, Zixuan Ling, Jianhao Zhu, Cenyuan Zhang, Xiaoqing Zheng, Xuanjing Huang

Aligning large language models (LLMs) with human preferences is crucial for enhancing their utility in terms of helpfulness, truthfulness, safety, harmlessness, and interestingness.

Tailoring Personality Traits in Large Language Models via Unsupervisedly-Built Personalized Lexicons

no code implementations25 Oct 2023 Tianlong Li, Shihan Dou, Changze Lv, Wenhao Liu, Jianhan Xu, Muling Wu, Zixuan Ling, Xiaoqing Zheng, Xuanjing Huang

Users can utilize UBPL to adjust the probability vectors of predicted words in the decoding phase of LLMs, thus influencing the personality expression of LLMs.

SpikeCLIP: A Contrastive Language-Image Pretrained Spiking Neural Network

no code implementations10 Oct 2023 Tianlong Li, Wenhao Liu, Changze Lv, Jianhan Xu, Cenyuan Zhang, Muling Wu, Xiaoqing Zheng, Xuanjing Huang

Spiking neural networks (SNNs) have demonstrated the capability to achieve comparable performance to deep neural networks (DNNs) in both visual and linguistic domains while offering the advantages of improved energy efficiency and adherence to biological plausibility.

Image Classification

Efficiently Aligned Cross-Lingual Transfer Learning for Conversational Tasks using Prompt-Tuning

1 code implementation3 Apr 2023 Lifu Tu, Jin Qu, Semih Yavuz, Shafiq Joty, Wenhao Liu, Caiming Xiong, Yingbo Zhou

Our results demonstrate the strong and efficient modeling ability of NLI-based classifiers and the large cross-lingual transfer improvements achieved by our aligned prompts, particularly in few-shot settings.

Cross-Lingual Transfer intent-classification +4

Near-Negative Distinction: Giving a Second Life to Human Evaluation Datasets

1 code implementation13 May 2022 Philippe Laban, Chien-Sheng Wu, Wenhao Liu, Caiming Xiong

Precisely assessing the progress in natural language generation (NLG) tasks is challenging, and human evaluation to establish a preference in a model's output over another is often necessary.

nlg evaluation Question Answering +3

A Generative Language Model for Few-shot Aspect-Based Sentiment Analysis

1 code implementation Findings (NAACL) 2022 Ehsan Hosseini-Asl, Wenhao Liu, Caiming Xiong

Our evaluation results on the single-task polarity prediction show that our approach outperforms the previous state-of-the-art (based on BERT) on average performance by a large margins in few-shot and full-shot settings.

Aspect-Based Sentiment Analysis Aspect-Based Sentiment Analysis (ABSA) +4

Structure Extraction in Task-Oriented Dialogues with Slot Clustering

2 code implementations28 Feb 2022 Liang Qiu, Chien-Sheng Wu, Wenhao Liu, Caiming Xiong

Extracting structure information from dialogue data can help us better understand user and system behaviors.

Clustering Data Augmentation +1

Exploring Neural Models for Query-Focused Summarization

1 code implementation Findings (NAACL) 2022 Jesse Vig, Alexander R. Fabbri, Wojciech Kryściński, Chien-Sheng Wu, Wenhao Liu

Query-focused summarization (QFS) aims to produce summaries that answer particular questions of interest, enabling greater user control and personalization.

Query-focused Summarization Transfer Learning

Open Vocabulary Object Detection with Pseudo Bounding-Box Labels

1 code implementation18 Nov 2021 Mingfei Gao, Chen Xing, Juan Carlos Niebles, Junnan Li, ran Xu, Wenhao Liu, Caiming Xiong

To enlarge the set of base classes, we propose a method to automatically generate pseudo bounding-box annotations of diverse objects from large-scale image-caption pairs.

Object object-detection +1

CaPE: Contrastive Parameter Ensembling for Reducing Hallucination in Abstractive Summarization

no code implementations14 Oct 2021 Prafulla Kumar Choubey, Alexander R. Fabbri, Jesse Vig, Chien-Sheng Wu, Wenhao Liu, Nazneen Fatema Rajani

Then, we fine-tune a base summarization model, which is trained on all training samples, on the clean (noisy) subset to obtain an \textit{expert} (\textit{anti-expert}) model.

Abstractive Text Summarization Hallucination +1

Improving Gender Fairness of Pre-Trained Language Models without Catastrophic Forgetting

no code implementations11 Oct 2021 Zahra Fatemi, Chen Xing, Wenhao Liu, Caiming Xiong

In this work, we empirically show that catastrophic forgetting occurs in such methods by evaluating them with general NLP tasks in GLUE.

coreference-resolution Fairness

HydraSum: Disentangling Stylistic Features in Text Summarization using Multi-Decoder Models

1 code implementation8 Oct 2021 Tanya Goyal, Nazneen Fatema Rajani, Wenhao Liu, Wojciech Kryściński

Summarization systems make numerous "decisions" about summary properties during inference, e. g. degree of copying, specificity and length of outputs, etc.

Abstractive Text Summarization Specificity

HydraSum - Disentangling Stylistic Features in Text Summarization using Multi-Decoder Models

no code implementations29 Sep 2021 Tanya Goyal, Nazneen Rajani, Wenhao Liu, Wojciech Maciej Kryscinski

Existing abstractive summarization models lack explicit control mechanisms that would allow users to influence the stylistic features of the model outputs.

Abstractive Text Summarization Specificity

Don’t throw away that linear head: Few-shot protein fitness prediction with generative models

no code implementations29 Sep 2021 Ben Krause, Nikhil Naik, Wenhao Liu, Ali Madani

Predicting the fitness, i. e. functional value, of a protein sequence is an important and challenging task in biology, particularly due to the scarcity of assay-labeled data.

Transfer Learning

QAConv: Question Answering on Informative Conversations

1 code implementation ACL 2022 Chien-Sheng Wu, Andrea Madotto, Wenhao Liu, Pascale Fung, Caiming Xiong

This paper introduces QAConv, a new question answering (QA) dataset that uses conversations as a knowledge source.

Question Answering

Learning from Mistakes: Using Mis-predictions as Harm Alerts in Language Pre-Training

no code implementations16 Dec 2020 Chen Xing, Wenhao Liu, Caiming Xiong

According to recent studies and our empirical observations, one possible reason is that some easy-to-fit patterns in the training data, such as frequently co-occurring word combinations, dominate and harm pre-training, making it hard for the model to fit more complex information.

Sentence

Cannot find the paper you are looking for? You can Submit a new open access paper.