Zero-Shot Text Classification
27 papers with code • 3 benchmarks • 4 datasets
Most implemented papers
Making Pre-trained Language Models Better Few-shot Learners
We present LM-BFF--better few-shot fine-tuning of language models--a suite of simple and complementary techniques for fine-tuning language models on a small number of annotated examples.
Benchmarking Zero-shot Text Classification: Datasets, Evaluation and Entailment Approach
0Shot-TC aims to associate an appropriate label with a piece of text, irrespective of the text domain and the aspect (e. g., topic, emotion, event, etc.)
Integrating Semantic Knowledge to Tackle Zero-shot Text Classification
Insufficient or even unavailable training data of emerging classes is a big challenge of many classification tasks, including text classification.
Text Classification Using Label Names Only: A Language Model Self-Training Approach
In this paper, we explore the potential of only using the label name of each class to train classification models on unlabeled data, without using any labeled documents.
Decoupling Knowledge from Memorization: Retrieval-augmented Prompt Learning
Specifically, vanilla prompt learning may struggle to utilize atypical instances by rote during fully-supervised training or overfit shallow patterns with low-shot data.
Evaluating Unsupervised Text Classification: Zero-shot and Similarity-based Approaches
Text classification of unseen classes is a challenging Natural Language Processing task and is mainly attempted using two different types of approaches.
Unsupervised Label Refinement Improves Dataless Text Classification
This reliance causes dataless classifiers to be highly sensitive to the choice of label descriptions and hinders the broader application of dataless classification in practice.
Issues with Entailment-based Zero-shot Text Classification
The general format of natural language inference (NLI) makes it tempting to be used for zero-shot text classification by casting any target label into a sentence of hypothesis and verifying whether or not it could be entailed by the input, aiming at generic classification applicable on any specified label space.
NSP-BERT: A Prompt-based Few-Shot Learner Through an Original Pre-training Task--Next Sentence Prediction
Using prompts to utilize language models to perform various downstream tasks, also known as prompt-based learning or prompt-learning, has lately gained significant success in comparison to the pre-train and fine-tune paradigm.
Generating Training Data with Language Models: Towards Zero-Shot Language Understanding
Pretrained language models (PLMs) have demonstrated remarkable performance in various natural language processing tasks: Unidirectional PLMs (e. g., GPT) are well known for their superior text generation capabilities; bidirectional PLMs (e. g., BERT) have been the prominent choice for natural language understanding (NLU) tasks.