Search Results for author: Qinyuan Ye

Found 14 papers, 10 papers with code

Prompt Engineering a Prompt Engineer

no code implementations9 Nov 2023 Qinyuan Ye, Maxamed Axmed, Reid Pryzant, Fereshte Khani

While recent works indicate that large language models can be meta-prompted to perform automatic prompt engineering, we argue that their potential is limited due to insufficient guidance for complex reasoning in the meta-prompt.

counterfactual Counterfactual Reasoning +2

How Predictable Are Large Language Model Capabilities? A Case Study on BIG-bench

1 code implementation24 May 2023 Qinyuan Ye, Harvey Yiyun Fu, Xiang Ren, Robin Jia

We investigate the predictability of large language model (LLM) capabilities: given records of past experiments using different model families, numbers of parameters, tasks, and numbers of in-context examples, can we accurately predict LLM performance on new experiment configurations?

Language Modelling Large Language Model

Estimating Large Language Model Capabilities without Labeled Test Data

1 code implementation24 May 2023 Harvey Yiyun Fu, Qinyuan Ye, Albert Xu, Xiang Ren, Robin Jia

In this paper, we propose the task of ICL accuracy estimation, in which we predict the accuracy of an LLM when doing in-context learning on a new task given only unlabeled test data for that task.

In-Context Learning Language Modelling +1

Eliciting and Understanding Cross-Task Skills with Task-Level Mixture-of-Experts

1 code implementation25 May 2022 Qinyuan Ye, Juan Zha, Xiang Ren

Recent works suggest that transformer models are capable of multi-tasking on diverse NLP tasks and adapting to new tasks efficiently.

Multi-Task Learning World Knowledge +1

Sparse Distillation: Speeding Up Text Classification by Using Bigger Student Models

1 code implementation NAACL 2022 Qinyuan Ye, Madian Khabsa, Mike Lewis, Sinong Wang, Xiang Ren, Aaron Jaech

Distilling state-of-the-art transformer models into lightweight student models is an effective way to reduce computation cost at inference time.

Domain Generalization Privacy Preserving +4

CrossFit: A Few-shot Learning Challenge for Cross-task Generalization in NLP

3 code implementations EMNLP 2021 Qinyuan Ye, Bill Yuchen Lin, Xiang Ren

Humans can learn a new language task efficiently with only few examples, by leveraging their knowledge obtained when learning prior tasks.

Few-Shot Learning

Refining Language Models with Compositional Explanations

1 code implementation NeurIPS 2021 Huihan Yao, Ying Chen, Qinyuan Ye, Xisen Jin, Xiang Ren

However, such a regularization technique lacks flexibility and coverage, since only importance scores towards a pre-defined list of features are adjusted, while more complex human knowledge such as feature interaction and pattern generalization can hardly be incorporated.

Fairness Language Modelling +2

Learning to Generate Task-Specific Adapters from Task Description

1 code implementation ACL 2021 Qinyuan Ye, Xiang Ren

Recent study further shows that they can learn to generalize to novel tasks, by including task descriptions as part of the source sequence and training the model with (source, target) examples.

Text Generation Zero-Shot Learning

Teaching Machine Comprehension with Compositional Explanations

2 code implementations Findings of the Association for Computational Linguistics 2020 Qinyuan Ye, Xiao Huang, Elizabeth Boschee, Xiang Ren

Advances in machine reading comprehension (MRC) rely heavily on the collection of large scale human-annotated examples in the form of (question, paragraph, answer) triples.

Data Augmentation Machine Reading Comprehension +1

Learning from Explanations with Neural Execution Tree

1 code implementation ICLR 2020 Ziqi Wang, Yujia Qin, Wenxuan Zhou, Jun Yan, Qinyuan Ye, Leonardo Neves, Zhiyuan Liu, Xiang Ren

While deep neural networks have achieved impressive performance on a range of NLP tasks, these data-hungry models heavily rely on labeled data, which restricts their applications in scenarios where data annotation is expensive.

Data Augmentation Multi-hop Question Answering +6

Looking Beyond Label Noise: Shifted Label Distribution Matters in Distantly Supervised Relation Extraction

1 code implementation IJCNLP 2019 Qinyuan Ye, Liyuan Liu, Maosen Zhang, Xiang Ren

In this paper, we study the problem what limits the performance of DS-trained neural models, conduct thorough analyses, and identify a factor that can influence the performance greatly, shifted label distribution.

Relation Relation Extraction

Cannot find the paper you are looking for? You can Submit a new open access paper.