Search Results for author: He He

Found 50 papers, 24 papers with code

Robustness and Adversarial Examples in Natural Language Processing

no code implementations EMNLP (ACL) 2021 Kai-Wei Chang, He He, Robin Jia, Sameer Singh

In particular, we will review recent studies on analyzing the weakness of NLP systems when facing adversarial inputs and data with a distribution shift.

Reward Gaming in Conditional Text Generation

no code implementations16 Nov 2022 Richard Yuanzhe Pang, Vishakh Padmakumar, Thibault Sellam, Ankur P. Parikh, He He

To align conditional text generation model outputs with desired behaviors, there has been an increasing focus on training the model using reinforcement learning (RL) with reward functions learned from human annotations.

Conditional Text Generation

Are All Spurious Features in Natural Language Alike? An Analysis through a Causal Lens

no code implementations25 Oct 2022 Nitish Joshi, Xiang Pan, He He

In case (i), we want the model to be invariant to the feature, which is neither necessary nor sufficient for prediction.

Help me write a poem: Instruction Tuning as a Vehicle for Collaborative Poetry Writing

1 code implementation25 Oct 2022 Tuhin Chakrabarty, Vishakh Padmakumar, He He

The core component of our system is a language model fine-tuned on a diverse collection of instructions for poetry writing.

Language Modelling

Robustification of Multilingual Language Models to Real-world Noise with Robust Contrastive Pretraining

no code implementations10 Oct 2022 Asa Cooper Stickland, Sailik Sengupta, Jason Krone, Saab Mansour, He He

Thus, to benchmark the performance of pretrained multilingual models, we construct noisy datasets covering five languages and four NLP tasks.

Data Augmentation

Nuisances via Negativa: Adjusting for Spurious Correlations via Data Augmentation

no code implementations4 Oct 2022 Aahlad Puli, Nitish Joshi, He He, Rajesh Ranganath

There exist features that are related to the label in the same way across different settings for that task; these are semantic features or semantics.

Data Augmentation Natural Language Inference

Language Models Are Greedy Reasoners: A Systematic Formal Analysis of Chain-of-Thought

1 code implementation3 Oct 2022 Abulhair Saparov, He He

Large language models (LLMs) have shown remarkable reasoning capabilities given chain-of-thought prompts (examples with intermediate reasoning steps).

Mathematical Reasoning Question Answering

On the Relation between Sensitivity and Accuracy in In-context Learning

no code implementations16 Sep 2022 Yanda Chen, Chen Zhao, Zhou Yu, Kathleen McKeown, He He

In-context learning (ICL) suffers from oversensitivity to the prompt, which makes it unreliable in real-world scenarios.

Exploring the Role of Task Transferability in Large-Scale Multi-Task Learning

no code implementations NAACL 2022 Vishakh Padmakumar, Leonard Lausen, Miguel Ballesteros, Sheng Zha, He He, George Karypis

Recent work has found that multi-task training with a large number of diverse tasks can uniformly improve downstream performance on unseen target tasks.

Multi-Task Learning Representation Learning

Amortized Noisy Channel Neural Machine Translation

no code implementations16 Dec 2021 Richard Yuanzhe Pang, He He, Kyunghyun Cho

For all three approaches, the generated translations fail to achieve rewards comparable to BSR, but the translation quality approximated by BLEU and BLEURT is similar to the quality of BSR-produced translations.

Imitation Learning Knowledge Distillation +4

QuALITY: Question Answering with Long Input Texts, Yes!

2 code implementations NAACL 2022 Richard Yuanzhe Pang, Alicia Parrish, Nitish Joshi, Nikita Nangia, Jason Phang, Angelica Chen, Vishakh Padmakumar, Johnny Ma, Jana Thompson, He He, Samuel R. Bowman

To enable building and testing models on long-document comprehension, we introduce QuALITY, a multiple-choice QA dataset with context passages in English that have an average length of about 5, 000 tokens, much longer than typical current models can process.

Multiple-choice Multiple Choice Question Answering (MCQA)

Machine-in-the-Loop Rewriting for Creative Image Captioning

1 code implementation NAACL 2022 Vishakh Padmakumar, He He

Machine-in-the-loop writing aims to enable humans to collaborate with models to complete their writing tasks more effectively.

Image Captioning Language Modelling

SeqPATE: Differentially Private Text Generation via Knowledge Distillation

no code implementations29 Sep 2021 Zhiliang Tian, Yingxiu Zhao, Ziyue Huang, Yu-Xiang Wang, Nevin Zhang, He He

Differentially private (DP) learning algorithms provide guarantees on identifying the existence of a training sample from model outputs.

Knowledge Distillation Sentence Completion +1

Types of Out-of-Distribution Texts and How to Detect Them

1 code implementation EMNLP 2021 Udit Arora, William Huang, He He

Despite agreement on the importance of detecting out-of-distribution (OOD) examples, there is little consensus on the formal definition of OOD examples and how to best detect them.

Density Estimation Language Modelling +2

An Investigation of the (In)effectiveness of Counterfactually Augmented Data

1 code implementation ACL 2022 Nitish Joshi, He He

While pretrained language models achieve excellent performance on natural language understanding benchmarks, they tend to rely on spurious correlations and generalize poorly to out-of-distribution (OOD) data.

Natural Language Understanding Pretrained Language Models

Unsupervised Extractive Summarization using Pointwise Mutual Information

1 code implementation EACL 2021 Vishakh Padmakumar, He He

Unsupervised approaches to extractive summarization usually rely on a notion of sentence importance defined by the semantic similarity between a sentence and the document.

Extractive Summarization Language Modelling +3

Text Generation by Learning from Demonstrations

1 code implementation ICLR 2021 Richard Yuanzhe Pang, He He

Current approaches to text generation largely rely on autoregressive models and maximum likelihood estimation.

Machine Translation Question Generation +3

An Empirical Study on Robustness to Spurious Correlations using Pre-trained Language Models

1 code implementation14 Jul 2020 Lifu Tu, Garima Lalwani, Spandana Gella, He He

Recent work has shown that pre-trained language models such as BERT improve robustness to spurious correlations in the dataset.

Multi-Task Learning Natural Language Inference +1

Different Set Domain Adaptation for Brain-Computer Interfaces: A Label Alignment Approach

1 code implementation3 Dec 2019 He He, Dongrui Wu

Currently, most domain adaptation approaches require the source domains to have the same feature space and label space as the target domain, which limits their applications, as the auxiliary data may have different feature spaces and/or different label spaces.

Domain Adaptation

A Dynamic Strategy Coach for Effective Negotiation

no code implementations WS 2019 Yiheng Zhou, He He, Alan W. black, Yulia Tsvetkov

We consider a bargaining scenario where a seller and a buyer negotiate the price of an item for sale through a text-based dialog.

Decision Making Text Generation

Multi-View Broad Learning System for Primate Oculomotor Decision Decoding

1 code implementation16 Aug 2019 Zhenhua Shi, Xiaomo Chen, Changming Zhao, He He, Veit Stuphorn, Dongrui Wu

Multi-view learning improves the learning performance by utilizing multi-view data: data collected from multiple sources, or feature sets extracted from the same data source.


GluonCV and GluonNLP: Deep Learning in Computer Vision and Natural Language Processing

4 code implementations9 Jul 2019 Jian Guo, He He, Tong He, Leonard Lausen, Mu Li, Haibin Lin, Xingjian Shi, Chenguang Wang, Junyuan Xie, Sheng Zha, Aston Zhang, Hang Zhang, Zhi Zhang, Zhongyue Zhang, Shuai Zheng, Yi Zhu

We present GluonCV and GluonNLP, the deep learning toolkits for computer vision and natural language processing based on Apache MXNet (incubating).

Pun Generation with Surprise

2 code implementations NAACL 2019 He He, Nanyun Peng, Percy Liang

We tackle the problem of generating a pun sentence given a pair of homophones (e. g., "died" and "dyed").

Association Language Modelling +1

Quizbowl: The Case for Incremental Question Answering

no code implementations9 Apr 2019 Pedro Rodriguez, Shi Feng, Mohit Iyyer, He He, Jordan Boyd-Graber

Throughout this paper, we show that collaborations with the vibrant trivia community have contributed to the quality of our dataset, spawned new research directions, and doubled as an exciting way to engage the public with research in machine learning and natural language processing.

BIG-bench Machine Learning Decision Making +1

QuAC: Question Answering in Context

no code implementations EMNLP 2018 Eunsol Choi, He He, Mohit Iyyer, Mark Yatskar, Wen-tau Yih, Yejin Choi, Percy Liang, Luke Zettlemoyer

We present QuAC, a dataset for Question Answering in Context that contains 14K information-seeking QA dialogs (100K questions in total).

Question Answering Reading Comprehension

QuAC : Question Answering in Context

no code implementations21 Aug 2018 Eunsol Choi, He He, Mohit Iyyer, Mark Yatskar, Wen-tau Yih, Yejin Choi, Percy Liang, Luke Zettlemoyer

We present QuAC, a dataset for Question Answering in Context that contains 14K information-seeking QA dialogs (100K questions in total).

Question Answering Reading Comprehension

Transfer Learning for Brain-Computer Interfaces: A Euclidean Space Data Alignment Approach

1 code implementation8 Aug 2018 He He, Dongrui Wu

Our approach has three desirable properties: 1) it aligns the EEG trials directly in the Euclidean space, and any signal processing, feature extraction and machine learning algorithms can then be applied to the aligned trials; 2) its computational cost is very low; and, 3) it is unsupervised and does not need any label information from the new subject.

EEG General Classification +1

Transfer Learning Enhanced Common Spatial Pattern Filtering for Brain Computer Interfaces (BCIs): Overview and a New Approach

no code implementations8 Aug 2018 He He, Dongrui Wu

The electroencephalogram (EEG) is the most widely used input for brain computer interfaces (BCIs), and common spatial pattern (CSP) is frequently used to spatially filter it to increase its signal-to-noise ratio.

EEG General Classification +1

Sharp Nearby, Fuzzy Far Away: How Neural Language Models Use Context

1 code implementation ACL 2018 Urvashi Khandelwal, He He, Peng Qi, Dan Jurafsky

We know very little about how neural language models (LM) use prior linguistic context.

Delete, Retrieve, Generate: A Simple Approach to Sentiment and Style Transfer

6 code implementations NAACL 2018 Juncen Li, Robin Jia, He He, Percy Liang

We consider the task of text attribute transfer: transforming a sentence to alter a specific attribute (e. g., sentiment) while preserving its attribute-independent content (e. g., changing "screen is just the right size" to "screen is too small").

Image Captioning Style Transfer +2

Learning Symmetric Collaborative Dialogue Agents with Dynamic Knowledge Graph Embeddings

2 code implementations ACL 2017 He He, Anusha Balakrishnan, Mihail Eric, Percy Liang

To model both structured knowledge and unstructured language, we propose a neural model with dynamic knowledge graph embeddings that evolve as the dialogue progresses.

Knowledge Graph Embeddings

Opponent Modeling in Deep Reinforcement Learning

1 code implementation18 Sep 2016 He He, Jordan Boyd-Graber, Kevin Kwok, Hal Daumé III

Opponent modeling is necessary in multi-agent settings where secondary agents with competing goals also adapt their strategies, yet it remains challenging because strategies interact with each other and change.

reinforcement-learning reinforcement Learning

Active Information Acquisition

no code implementations5 Feb 2016 He He, Paul Mineiro, Nikos Karampatziakis

We propose a general framework for sequential and dynamic acquisition of useful information in order to solve a particular task.

General Reinforcement Learning reinforcement Learning +1

Learning to Search for Dependencies

no code implementations18 Mar 2015 Kai-Wei Chang, He He, Hal Daumé III, John Langford

We demonstrate that a dependency parser can be built using a credit assignment compiler which removes the burden of worrying about low-level machine learning details from the parser implementation.

BIG-bench Machine Learning

Learning to Search in Branch and Bound Algorithms

no code implementations NeurIPS 2014 He He, Hal Daume III, Jason M. Eisner

Branch-and-bound is a widely used method in combinatorial optimization, including mixed integer programming, structured prediction and MAP inference.

Combinatorial Optimization Imitation Learning +1

A Credit Assignment Compiler for Joint Prediction

no code implementations NeurIPS 2016 Kai-Wei Chang, He He, Hal Daumé III, John Langford, Stephane Ross

Many machine learning applications involve jointly predicting multiple mutually dependent output variables.

Imitation Learning by Coaching

no code implementations NeurIPS 2012 He He, Jason Eisner, Hal Daume

However, it is important to note that these guarantees depend on how well the policy we found can imitate the oracle on the training data.

Imitation Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.