Search Results for author: Patrick Xia

Found 30 papers, 13 papers with code

Generative Adapter: Contextualizing Language Models in Parameters with A Single Forward Pass

no code implementations8 Nov 2024 Tong Chen, Hao Fang, Patrick Xia, Xiaodong Liu, Benjamin Van Durme, Luke Zettlemoyer, Jianfeng Gao, Hao Cheng

In StreamingQA, our approach is effective in injecting knowledge into the LM's parameters, achieving a 63. 5% improvement in F1 score over the model with supervised fine-tuning (from $19. 5$ to $31. 5$) for contexts as long as 32K tokens.

In-Context Learning Self-Supervised Learning

Multi-Field Adaptive Retrieval

no code implementations26 Oct 2024 Millicent Li, Tongfei Chen, Benjamin Van Durme, Patrick Xia

Document retrieval for tasks such as search and retrieval-augmented generation typically involves datasets that are unstructured: free-form text without explicit internal structure in each document.

Document Ranking Retrieval

SMART: Self-learning Meta-strategy Agent for Reasoning Tasks

1 code implementation21 Oct 2024 Rongxing Liu, Kumar Shridhar, Manish Prajapat, Patrick Xia, Mrinmaya Sachan

This inefficiency raises the question: Can LMs learn to select the optimal strategy in the first attempt, without a need for refinement?

GSM8K Self-Learning

Learning to Retrieve Iteratively for In-Context Learning

no code implementations20 Jun 2024 Yunmo Chen, Tongfei Chen, Harsh Jhamtani, Patrick Xia, Richard Shin, Jason Eisner, Benjamin Van Durme

We introduce iterative retrieval, a novel framework that empowers retrievers to make iterative decisions through policy optimization.

Combinatorial Optimization In-Context Learning +2

Interpreting User Requests in the Context of Natural Language Standing Instructions

1 code implementation16 Nov 2023 Nikita Moghe, Patrick Xia, Jacob Andreas, Jason Eisner, Benjamin Van Durme, Harsh Jhamtani

Users of natural language interfaces, generally powered by Large Language Models (LLMs), often must repeat their preferences each time they make a similar request.

Few-Shot Adaptation for Parsing Contextual Utterances with LLMs

1 code implementation18 Sep 2023 Kevin Lin, Patrick Xia, Hao Fang

We evaluate the ability of semantic parsers based on large language models (LLMs) to handle contextual utterances.

In-Context Learning Semantic Parsing

Natural Language Decomposition and Interpretation of Complex Utterances

no code implementations15 May 2023 Harsh Jhamtani, Hao Fang, Patrick Xia, Eran Levy, Jacob Andreas, Ben Van Durme

Designing natural language interfaces has historically required collecting supervised data to translate user requests into carefully designed intent representations.

Language Modeling Language Modelling

Automatic Document Selection for Efficient Encoder Pretraining

no code implementations20 Oct 2022 Yukun Feng, Patrick Xia, Benjamin Van Durme, João Sedoc

Building pretrained language models is considered expensive and data-intensive, but must we increase dataset size to achieve better performance?

Sentence

Multilingual Coreference Resolution in Multiparty Dialogue

1 code implementation2 Aug 2022 Boyuan Zheng, Patrick Xia, Mahsa Yarmohammadi, Benjamin Van Durme

Existing multiparty dialogue datasets for entity coreference resolution are nascent, and many challenges are still unaddressed.

coreference-resolution Data Augmentation

Pruning Pretrained Encoders with a Multitask Objective

no code implementations10 Dec 2021 Patrick Xia, Richard Shin

The sizes of pretrained language models make them challenging and expensive to use when there are multiple desired downstream tasks.

On Generalization in Coreference Resolution

2 code implementations CRAC (ACL) 2021 Shubham Toshniwal, Patrick Xia, Sam Wiseman, Karen Livescu, Kevin Gimpel

While coreference resolution is defined independently of dataset domain, most models for performing coreference resolution do not transfer well to unseen domains.

coreference-resolution Data Augmentation

Moving on from OntoNotes: Coreference Resolution Model Transfer

2 code implementations EMNLP 2021 Patrick Xia, Benjamin Van Durme

Academic neural models for coreference resolution (coref) are typically trained on a single dataset, OntoNotes, and model improvements are benchmarked on that same dataset.

coreference-resolution model

Incremental Neural Coreference Resolution in Constant Memory

1 code implementation EMNLP 2020 Patrick Xia, João Sedoc, Benjamin Van Durme

We investigate modeling coreference resolution under a fixed memory constraint by extending an incremental clustering algorithm to utilize contextualized encoders and neural components.

Clustering coreference-resolution +1

Multi-Sentence Argument Linking

no code implementations ACL 2020 Seth Ebner, Patrick Xia, Ryan Culkin, Kyle Rawlins, Benjamin Van Durme

We present a novel document-level model for finding argument spans that fill an event's roles, connecting related ideas in sentence-level semantic role labeling and coreference resolution.

coreference-resolution Semantic Role Labeling +2

Improved Lexically Constrained Decoding for Translation and Monolingual Rewriting

1 code implementation NAACL 2019 J. Edward Hu, Huda Khayrallah, Ryan Culkin, Patrick Xia, Tongfei Chen, Matt Post, Benjamin Van Durme

Lexically-constrained sequence decoding allows for explicit positive or negative phrase-based constraints to be placed on target output strings in generation tasks such as machine translation or monolingual text rewriting.

Data Augmentation Machine Translation +3

Looking for ELMo's friends: Sentence-Level Pretraining Beyond Language Modeling

no code implementations ICLR 2019 Samuel R. Bowman, Ellie Pavlick, Edouard Grave, Benjamin Van Durme, Alex Wang, Jan Hula, Patrick Xia, Raghavendra Pappagari, R. Thomas McCoy, Roma Patel, Najoung Kim, Ian Tenney, Yinghui Huang, Katherin Yu, Shuning Jin, Berlin Chen

Work on the problem of contextualized word representation—the development of reusable neural network components for sentence understanding—has recently seen a surge of progress centered on the unsupervised pretraining task of language modeling with methods like ELMo (Peters et al., 2018).

Language Modeling Language Modelling +1

Probing What Different NLP Tasks Teach Machines about Function Word Comprehension

no code implementations SEMEVAL 2019 Najoung Kim, Roma Patel, Adam Poliak, Alex Wang, Patrick Xia, R. Thomas McCoy, Ian Tenney, Alexis Ross, Tal Linzen, Benjamin Van Durme, Samuel R. Bowman, Ellie Pavlick

Our results show that pretraining on language modeling performs the best on average across our probing tasks, supporting its widespread use for pretraining state-of-the-art NLP models, and CCG supertagging and NLI pretraining perform comparably.

CCG Supertagging Language Modeling +4

Annotating Character Relationships in Literary Texts

no code implementations2 Dec 2015 Philip Massey, Patrick Xia, David Bamman, Noah A. Smith

We present a dataset of manually annotated relationships between characters in literary texts, in order to support the training and evaluation of automatic methods for relation type prediction in this domain (Makazhanov et al., 2014; Kokkinakis, 2013) and the broader computational analysis of literary character (Elson et al., 2010; Bamman et al., 2014; Vala et al., 2015; Flekova and Gurevych, 2015).

Type prediction

Cannot find the paper you are looking for? You can Submit a new open access paper.