Search Results for author: Jesse Vig

Found 19 papers, 11 papers with code

Beyond the Chat: Executable and Verifiable Text-Editing with LLMs

no code implementations27 Sep 2023 Philippe Laban, Jesse Vig, Marti A. Hearst, Caiming Xiong, Chien-Sheng Wu

Conversational interfaces powered by Large Language Models (LLMs) have recently become a popular way to obtain feedback during document editing.

XGen-7B Technical Report

1 code implementation7 Sep 2023 Erik Nijkamp, Tian Xie, Hiroaki Hayashi, Bo Pang, Congying Xia, Chen Xing, Jesse Vig, Semih Yavuz, Philippe Laban, Ben Krause, Senthil Purushwalkam, Tong Niu, Wojciech Kryściński, Lidiya Murakhovs'ka, Prafulla Kumar Choubey, Alex Fabbri, Ye Liu, Rui Meng, Lifu Tu, Meghana Bhat, Chien-Sheng Wu, Silvio Savarese, Yingbo Zhou, Shafiq Joty, Caiming Xiong

Most open-source LLMs, on the other hand, are limited in their ability to support longer sequence lengths, which is a key requirement for many tasks that require inference over an input context.

2k 8k

Did You Read the Instructions? Rethinking the Effectiveness of Task Definitions in Instruction Learning

1 code implementation1 Jun 2023 Fan Yin, Jesse Vig, Philippe Laban, Shafiq Joty, Caiming Xiong, Chien-Sheng Jason Wu

Large language models (LLMs) have shown impressive performance in following natural language instructions to solve unseen tasks.

SWiPE: A Dataset for Document-Level Simplification of Wikipedia Pages

1 code implementation30 May 2023 Philippe Laban, Jesse Vig, Wojciech Kryscinski, Shafiq Joty, Caiming Xiong, Chien-Sheng Wu

Text simplification research has mostly focused on sentence-level simplification, even though many desirable edits - such as adding relevant background information or reordering content - may require document-level context.

Sentence Text Simplification

Improving Factual Consistency in Summarization with Compression-Based Post-Editing

1 code implementation11 Nov 2022 Alexander R. Fabbri, Prafulla Kumar Choubey, Jesse Vig, Chien-Sheng Wu, Caiming Xiong

We propose to use sentence-compression data to train the post-editing model to take a summary with extrinsic entity errors marked with special tokens and output a compressed, well-formed summary with those errors removed.

Informativeness Sentence +1

Interactive Model Cards: A Human-Centered Approach to Model Documentation

no code implementations5 May 2022 Anamaria Crisan, Margaret Drouhard, Jesse Vig, Nazneen Rajani

Deep learning models for natural language processing (NLP) are increasingly adopted and deployed by analysts without formal training in NLP or machine learning (ML).

Ethics

iSEA: An Interactive Pipeline for Semantic Error Analysis of NLP Models

1 code implementation8 Mar 2022 Jun Yuan, Jesse Vig, Nazneen Rajani

Error analysis in NLP models is essential to successful model development and deployment.

Exploring Neural Models for Query-Focused Summarization

1 code implementation Findings (NAACL) 2022 Jesse Vig, Alexander R. Fabbri, Wojciech Kryściński, Chien-Sheng Wu, Wenhao Liu

Query-focused summarization (QFS) aims to produce summaries that answer particular questions of interest, enabling greater user control and personalization.

Query-focused Summarization Transfer Learning

CaPE: Contrastive Parameter Ensembling for Reducing Hallucination in Abstractive Summarization

no code implementations14 Oct 2021 Prafulla Kumar Choubey, Alexander R. Fabbri, Jesse Vig, Chien-Sheng Wu, Wenhao Liu, Nazneen Fatema Rajani

Then, we fine-tune a base summarization model, which is trained on all training samples, on the clean (noisy) subset to obtain an \textit{expert} (\textit{anti-expert}) model.

Abstractive Text Summarization Hallucination +1

SummVis: Interactive Visual Analysis of Models, Data, and Evaluation for Text Summarization

2 code implementations ACL 2021 Jesse Vig, Wojciech Kryściński, Karan Goel, Nazneen Fatema Rajani

Novel neural architectures, training strategies, and the availability of large-scale corpora haven been the driving force behind recent progress in abstractive text summarization.

Abstractive Text Summarization

Robustness Gym: Unifying the NLP Evaluation Landscape

2 code implementations NAACL 2021 Karan Goel, Nazneen Rajani, Jesse Vig, Samson Tan, Jason Wu, Stephan Zheng, Caiming Xiong, Mohit Bansal, Christopher Ré

Despite impressive performance on standard benchmarks, deep neural networks are often brittle when deployed in real-world systems.

Entity Linking

Profile Prediction: An Alignment-Based Pre-Training Task for Protein Sequence Models

no code implementations1 Dec 2020 Pascal Sturmfels, Jesse Vig, Ali Madani, Nazneen Fatema Rajani

Recent deep-learning approaches to protein prediction have shown that pre-training on unlabeled data can yield useful representations for downstream tasks.

Language Modelling Masked Language Modeling +1

BERTology Meets Biology: Interpreting Attention in Protein Language Models

2 code implementations ICLR 2021 Jesse Vig, Ali Madani, Lav R. Varshney, Caiming Xiong, Richard Socher, Nazneen Fatema Rajani

Transformer architectures have proven to learn useful representations for protein classification and generation tasks.

Causal Mediation Analysis for Interpreting Neural NLP: The Case of Gender Bias

1 code implementation26 Apr 2020 Jesse Vig, Sebastian Gehrmann, Yonatan Belinkov, Sharon Qian, Daniel Nevo, Simas Sakenis, Jason Huang, Yaron Singer, Stuart Shieber

Common methods for interpreting neural models in natural language processing typically examine either their structure or their behavior, but not both.

A Multiscale Visualization of Attention in the Transformer Model

3 code implementations ACL 2019 Jesse Vig

The Transformer is a sequence model that forgoes traditional recurrent architectures in favor of a fully attention-based approach.

Analyzing the Structure of Attention in a Transformer Language Model

no code implementations WS 2019 Jesse Vig, Yonatan Belinkov

The Transformer is a fully attention-based alternative to recurrent networks that has achieved state-of-the-art results across a range of NLP tasks.

Language Modelling

Visualizing Attention in Transformer-Based Language Representation Models

no code implementations4 Apr 2019 Jesse Vig

We present an open-source tool for visualizing multi-head self-attention in Transformer-based language representation models.

Language Modelling

Cannot find the paper you are looking for? You can Submit a new open access paper.