Search Results for author: Arman Cohan

Found 87 papers, 54 papers with code

Zero- and Few-Shot NLP with Pretrained Language Models

no code implementations ACL 2022 Iz Beltagy, Arman Cohan, Robert Logan IV, Sewon Min, Sameer Singh

The ability to efficiently learn from little-to-no data is critical to applying NLP to tasks where data collection is costly or otherwise difficult.

Few-Shot Learning

NExT: Teaching Large Language Models to Reason about Code Execution

no code implementations23 Apr 2024 Ansong Ni, Miltiadis Allamanis, Arman Cohan, Yinlin Deng, Kensen Shi, Charles Sutton, Pengcheng Yin

A fundamental skill among human developers is the ability to understand and reason about program execution.

Program Repair

MIMIR: A Streamlined Platform for Personalized Agent Tuning in Domain Expertise

no code implementations3 Apr 2024 Chunyuan Deng, Xiangru Tang, Yilun Zhao, Hanming Wang, Haoran Wang, Wangchunshu Zhou, Arman Cohan, Mark Gerstein

Recently, large language models (LLMs) have evolved into interactive agents, proficient in planning, tool use, and task execution across a wide variety of tasks.

FollowIR: Evaluating and Teaching Information Retrieval Models to Follow Instructions

2 code implementations22 Mar 2024 Orion Weller, Benjamin Chang, Sean MacAvaney, Kyle Lo, Arman Cohan, Benjamin Van Durme, Dawn Lawrie, Luca Soldaini

We introduce our dataset FollowIR, which contains a rigorous instruction evaluation benchmark as well as a training set for helping IR models learn to better follow real-world instructions.

Information Retrieval Retrieval

On the Benefits of Fine-Grained Loss Truncation: A Case Study on Factuality in Summarization

1 code implementation9 Mar 2024 Lorenzo Jaime Yu Flores, Arman Cohan

We study the behavior of the underlying losses between factual and non-factual examples, to understand and refine the performance of LT. We demonstrate that LT's performance is limited when the underlying assumption that noisy targets have higher NLL loss is not satisfied, and find that word-level NLL among entities provides better signal for distinguishing factuality.

Hallucination Text Summarization

Quantifying Contamination in Evaluating Code Generation Capabilities of Language Models

1 code implementation6 Mar 2024 Martin Riddell, Ansong Ni, Arman Cohan

While large language models have achieved remarkable performance on various code generation benchmarks, there have been growing concerns regarding potential contamination of these benchmarks as they may be leaked into pretraining and finetuning data.

Code Generation Memorization +1

Calibrating Long-form Generations from Large Language Models

no code implementations9 Feb 2024 Yukun Huang, Yixin Liu, Raghuveer Thirukovalluru, Arman Cohan, Bhuwan Dhingra

Addressing this gap, we introduce a unified calibration framework, in which both the correctness of the LLMs' responses and their associated confidence levels are treated as distributions across a range of scores.

Prioritizing Safeguarding Over Autonomy: Risks of LLM Agents for Science

no code implementations6 Feb 2024 Xiangru Tang, Qiao Jin, Kunlun Zhu, Tongxin Yuan, Yichi Zhang, Wangchunshu Zhou, Meng Qu, Yilun Zhao, Jian Tang, Zhuosheng Zhang, Arman Cohan, Zhiyong Lu, Mark Gerstein

Intelligent agents powered by large language models (LLMs) have demonstrated substantial promise in autonomously conducting experiments and facilitating scientific discoveries across various disciplines.

Observable Propagation: A Data-Efficient Approach to Uncover Feature Vectors in Transformers

1 code implementation26 Dec 2023 Jacob Dunefsky, Arman Cohan

Our results suggest that ObsProp surpasses traditional approaches for finding feature vectors in the low-data regime, and that ObsProp can be used to better understand the mechanisms responsible for bias in large language models.

Investigating Data Contamination in Modern Benchmarks for Large Language Models

no code implementations16 Nov 2023 Chunyuan Deng, Yilun Zhao, Xiangru Tang, Mark Gerstein, Arman Cohan

Recent observations have underscored a disparity between the inflated benchmark scores and the actual performance of LLMs, raising concerns about potential contamination of evaluation benchmarks.

Common Sense Reasoning Multiple-choice +1

ML-Bench: Evaluating Large Language Models for Code Generation in Repository-Level Machine Learning Tasks

1 code implementation16 Nov 2023 Yuliang Liu, Xiangru Tang, Zefan Cai, Junjie Lu, Yichi Zhang, Yanjun Shao, Zexuan Deng, Helan Hu, Kaikai An, Ruijun Huang, Shuzheng Si, Sheng Chen, Haozhe Zhao, Liang Chen, Yan Wang, Tianyu Liu, Zhiwei Jiang, Baobao Chang, Yujia Qin, Wangchunshu Zhou, Yilun Zhao, Arman Cohan, Mark Gerstein

While Large Language Models (LLMs) have demonstrated proficiency in code generation benchmarks, translating these results into practical development scenarios - where leveraging existing repository-level libraries is the norm - remains challenging.

Code Generation Navigate

On Evaluating the Integration of Reasoning and Action in LLM Agents with Database Question Answering

no code implementations16 Nov 2023 Linyong Nan, Ellen Zhang, Weijin Zou, Yilun Zhao, Wenfei Zhou, Arman Cohan

A key discovery is the identification of two primary bottlenecks hindering effective interaction: the capacity for planning and the ability to generate multiple SQL queries.

Question Answering Retrieval

KnowledgeMath: Knowledge-Intensive Math Word Problem Solving in Finance Domains

1 code implementation16 Nov 2023 Yilun Zhao, Hongjun Liu, Yitao Long, Rui Zhang, Chen Zhao, Arman Cohan

We introduce KnowledgeMath, a novel benchmark designed to evaluate LLMs' capabilities in applying financial knowledge to solve complex math word problems.

Math Math Word Problem Solving +1

DocMath-Eval: Evaluating Numerical Reasoning Capabilities of LLMs in Understanding Long Documents with Tabular Data

no code implementations16 Nov 2023 Yilun Zhao, Yitao Long, Hongjun Liu, Linyong Nan, Lyuhao Chen, Ryo Kamoi, Yixin Liu, Xiangru Tang, Rui Zhang, Arman Cohan

This paper introduces DocMath-Eval, a comprehensive benchmark specifically designed to evaluate the numerical reasoning and problem-solving capabilities of LLMs in the context of understanding and analyzing financial documents containing both text and tables.

Math

MedAgents: Large Language Models as Collaborators for Zero-shot Medical Reasoning

1 code implementation16 Nov 2023 Xiangru Tang, Anni Zou, Zhuosheng Zhang, Ziming Li, Yilun Zhao, Xingyao Zhang, Arman Cohan, Mark Gerstein

Large language models (LLMs), despite their remarkable progress across various general domains, encounter significant barriers in medicine and healthcare.

Back to Basics: A Simple Recipe for Improving Out-of-Domain Retrieval in Dense Encoders

1 code implementation16 Nov 2023 Hyunji Lee, Luca Soldaini, Arman Cohan, Minjoon Seo, Kyle Lo

Prevailing research practice today often relies on training dense retrievers on existing large datasets such as MSMARCO and then experimenting with ways to improve zero-shot generalization capabilities to unseen domains.

Data Augmentation Domain Generalization +2

Benchmarking Generation and Evaluation Capabilities of Large Language Models for Instruction Controllable Summarization

1 code implementation15 Nov 2023 Yixin Liu, Alexander R. Fabbri, Jiawen Chen, Yilun Zhao, Simeng Han, Shafiq Joty, PengFei Liu, Dragomir Radev, Chien-Sheng Wu, Arman Cohan

Our study reveals that instruction controllable text summarization remains a challenging task for LLMs, since (1) all LLMs evaluated still make factual and other types of errors in their summaries; (2) all LLM-based evaluation methods cannot achieve a strong alignment with human annotators when judging the quality of candidate summaries; (3) different LLMs show large performance gaps in summary generation and evaluation.

Benchmarking Text Summarization

Medical Text Simplification: Optimizing for Readability with Unlikelihood Training and Reranked Beam Search Decoding

1 code implementation17 Oct 2023 Lorenzo Jaime Yu Flores, Heyuan Huang, Kejian Shi, Sophie Chheang, Arman Cohan

Text simplification has emerged as an increasingly useful application of AI for bridging the communication gap in specialized fields such as medicine, where the lexicon is often dominated by technical jargon and complex constructs.

Text Simplification

L2CEval: Evaluating Language-to-Code Generation Capabilities of Large Language Models

no code implementations29 Sep 2023 Ansong Ni, Pengcheng Yin, Yilun Zhao, Martin Riddell, Troy Feng, Rui Shen, Stephen Yin, Ye Liu, Semih Yavuz, Caiming Xiong, Shafiq Joty, Yingbo Zhou, Dragomir Radev, Arman Cohan

Recently, large language models (LLMs), especially those that are pretrained on code, have demonstrated strong capabilities in generating programs from natural language inputs in a few-shot or even zero-shot manner.

Code Generation Math +1

ODSum: New Benchmarks for Open Domain Multi-Document Summarization

1 code implementation16 Sep 2023 Yijie Zhou, Kejian Shi, Wencai Zhang, Yixin Liu, Yilun Zhao, Arman Cohan

Open-domain Multi-Document Summarization (ODMDS) is a critical tool for condensing vast arrays of documents into coherent, concise summaries.

Document Summarization Multi-Document Summarization +1

Struc-Bench: Are Large Language Models Really Good at Generating Complex Structured Data?

1 code implementation16 Sep 2023 Xiangru Tang, Yiming Zong, Jason Phang, Yilun Zhao, Wangchunshu Zhou, Arman Cohan, Mark Gerstein

Despite the remarkable capabilities of Large Language Models (LLMs) like GPT-4, producing complex, structured tabular data remains challenging.

Hallucination

Investigating Table-to-Text Generation Capabilities of LLMs in Real-World Information Seeking Scenarios

2 code implementations24 May 2023 Yilun Zhao, Haowei Zhang, Shengyun Si, Linyong Nan, Xiangru Tang, Arman Cohan

These include the LogicNLG and our newly-constructed LoTNLG datasets for data insight generation, along with the FeTaQA and our newly-constructed F2WTQ datasets for query-based generation.

Table-to-Text Generation

A Question Answering Framework for Decontextualizing User-facing Snippets from Scientific Documents

no code implementations24 May 2023 Benjamin Newman, Luca Soldaini, Raymond Fok, Arman Cohan, Kyle Lo

Many real-world applications (e. g., note taking, search) require extracting a sentence or paragraph from a document and showing that snippet to a human outside of the source document.

Question Answering Question Generation +2

Peek Across: Improving Multi-Document Modeling via Cross-Document Question-Answering

1 code implementation24 May 2023 Avi Caciularu, Matthew E. Peters, Jacob Goldberger, Ido Dagan, Arman Cohan

The integration of multi-document pre-training objectives into language models has resulted in remarkable improvements in multi-document downstream tasks.

Query-focused Summarization Question Answering +2

On Learning to Summarize with Large Language Models as References

1 code implementation23 May 2023 Yixin Liu, Kejian Shi, Katherine S He, Longtian Ye, Alexander R. Fabbri, PengFei Liu, Dragomir Radev, Arman Cohan

Meanwhile, we perform a meta-analysis on this new learning setting that reveals a discrepancy between human and LLM-based evaluation, highlighting the benefits and risks of this LLM-as-reference setting we investigated.

Contrastive Learning Text Summarization

QTSumm: Query-Focused Summarization over Tabular Data

2 code implementations23 May 2023 Yilun Zhao, Zhenting Qi, Linyong Nan, Boyu Mi, Yixin Liu, Weijin Zou, Simeng Han, Ruizhe Chen, Xiangru Tang, Yumo Xu, Dragomir Radev, Arman Cohan

Motivated by this, we define a new query-focused table summarization task, where text generation models have to perform human-like reasoning and analysis over the given table to generate a tailored summary.

Query-focused Summarization Table-to-Text Generation

Enhancing Few-shot Text-to-SQL Capabilities of Large Language Models: A Study on Prompt Design Strategies

no code implementations21 May 2023 Linyong Nan, Yilun Zhao, Weijin Zou, Narutatsu Ri, Jaesung Tae, Ellen Zhang, Arman Cohan, Dragomir Radev

In-context learning (ICL) has emerged as a new approach to various natural language processing tasks, utilizing large language models (LLMs) to make predictions based on context that has been supplemented with a few examples or task-specific instructions.

In-Context Learning Question Answering +1

Inference-time Re-ranker Relevance Feedback for Neural Information Retrieval

no code implementations19 May 2023 Revanth Gangi Reddy, Pradeep Dasigi, Md Arafat Sultan, Arman Cohan, Avirup Sil, Heng Ji, Hannaneh Hajishirzi

Neural information retrieval often adopts a retrieve-and-rerank framework: a bi-encoder network first retrieves K (e. g., 100) candidates that are then re-ranked using a more powerful cross-encoder model to rank the better candidates higher.

Information Retrieval Retrieval

LongEval: Guidelines for Human Evaluation of Faithfulness in Long-form Summarization

1 code implementation30 Jan 2023 Kalpesh Krishna, Erin Bransom, Bailey Kuehl, Mohit Iyyer, Pradeep Dasigi, Arman Cohan, Kyle Lo

Motivated by our survey, we present LongEval, a set of guidelines for human evaluation of faithfulness in long-form summaries that addresses the following challenges: (1) How can we achieve high inter-annotator agreement on faithfulness scores?

Open Domain Multi-document Summarization: A Comprehensive Study of Model Brittleness under Retrieval

no code implementations20 Dec 2022 John Giorgi, Luca Soldaini, Bo wang, Gary Bader, Kyle Lo, Lucy Lu Wang, Arman Cohan

Via extensive automatic and human evaluation, we determine: (1) state-of-the-art summarizers suffer large reductions in performance when applied to open-domain MDS, (2) additional training in the open-domain setting can reduce this sensitivity to imperfect retrieval, and (3) summarizers are insensitive to the retrieval of duplicate documents and the order of retrieved documents, but highly sensitive to other errors, like the retrieval of irrelevant documents.

Document Summarization Multi-Document Summarization +1

SciRepEval: A Multi-Format Benchmark for Scientific Document Representations

2 code implementations23 Nov 2022 Amanpreet Singh, Mike D'Arcy, Arman Cohan, Doug Downey, Sergey Feldman

In response, we introduce SciRepEval, the first comprehensive benchmark for training and evaluating scientific document representations.

SciFact-Open: Towards open-domain scientific claim verification

1 code implementation25 Oct 2022 David Wadden, Kyle Lo, Bailey Kuehl, Arman Cohan, Iz Beltagy, Lucy Lu Wang, Hannaneh Hajishirzi

While research on scientific claim verification has led to the development of powerful systems that appear to approach human performance, these approaches have yet to be tested in a realistic setting against large corpora of scientific literature.

Claim Verification Information Retrieval +1

Embedding Recycling for Language Models

1 code implementation11 Jul 2022 Jon Saad-Falcon, Amanpreet Singh, Luca Soldaini, Mike D'Arcy, Arman Cohan, Doug Downey

Real-world applications of neural language models often involve running many different models over the same corpus.

Question Answering Text Classification

Improving the Generalizability of Depression Detection by Leveraging Clinical Questionnaires

1 code implementation ACL 2022 Thong Nguyen, Andrew Yates, Ayah Zirikly, Bart Desmet, Arman Cohan

In dataset-transfer experiments on three social media datasets, we find that grounding the model in PHQ9's symptoms substantially improves its ability to generalize to out-of-distribution data compared to a standard BERT-based approach.

Depression Detection Domain Generalization

Generating Scientific Claims for Zero-Shot Scientific Fact Checking

1 code implementation ACL 2022 Dustin Wright, David Wadden, Kyle Lo, Bailey Kuehl, Arman Cohan, Isabelle Augenstein, Lucy Lu Wang

To address this challenge, we propose scientific claim generation, the task of generating one or more atomic and verifiable claims from scientific sentences, and demonstrate its usefulness in zero-shot fact checking for biomedical claims.

Fact Checking Negation

MultiVerS: Improving scientific claim verification with weak supervision and full-document context

3 code implementations Findings (NAACL) 2022 David Wadden, Kyle Lo, Lucy Lu Wang, Arman Cohan, Iz Beltagy, Hannaneh Hajishirzi

Our approach outperforms two competitive baselines on three scientific claim verification datasets, with particularly strong performance in zero / few-shot domain adaptation experiments.

Claim Verification Domain Adaptation +2

PRIMERA: Pyramid-based Masked Sentence Pre-training for Multi-document Summarization

2 code implementations ACL 2022 Wen Xiao, Iz Beltagy, Giuseppe Carenini, Arman Cohan

We introduce PRIMERA, a pre-trained model for multi-document representation with a focus on summarization that reduces the need for dataset-specific architectures and large amounts of fine-tuning labeled data.

Abstractive Text Summarization Document Summarization +2

FLEX: Unifying Evaluation for Few-Shot NLP

2 code implementations NeurIPS 2021 Jonathan Bragg, Arman Cohan, Kyle Lo, Iz Beltagy

Few-shot NLP research is highly active, yet conducted in disjoint research threads with evaluation suites that lack challenging-yet-realistic testing setups and fail to employ careful experimental design.

Few-Shot Learning Language Modelling

Beyond Paragraphs: NLP for Long Sequences

1 code implementation NAACL 2021 Iz Beltagy, Arman Cohan, Hannaneh Hajishirzi, Sewon Min, Matthew E. Peters

In this tutorial, we aim at bringing interested NLP researchers up to speed about the recent and ongoing techniques for document-level representation learning.

Representation Learning

CDLM: Cross-Document Language Modeling

2 code implementations Findings (EMNLP) 2021 Avi Caciularu, Arman Cohan, Iz Beltagy, Matthew E. Peters, Arie Cattan, Ido Dagan

We introduce a new pretraining approach geared for multi-document language modeling, incorporating two key ideas into the masked language modeling self-supervised objective.

Citation Recommendation Coreference Resolution +6

ABNIRML: Analyzing the Behavior of Neural IR Models

2 code implementations2 Nov 2020 Sean MacAvaney, Sergey Feldman, Nazli Goharian, Doug Downey, Arman Cohan

Pretrained contextualized language models such as BERT and T5 have established a new state-of-the-art for ad-hoc search.

Language Modelling Sentence

SLEDGE-Z: A Zero-Shot Baseline for COVID-19 Literature Search

no code implementations EMNLP 2020 Sean MacAvaney, Arman Cohan, Nazli Goharian

With worldwide concerns surrounding the Severe Acute Respiratory Syndrome Coronavirus 2 (SARS-CoV-2), there is a rapidly growing body of scientific literature on the virus.

Re-Ranking

SLEDGE: A Simple Yet Effective Baseline for COVID-19 Scientific Knowledge Search

1 code implementation5 May 2020 Sean MacAvaney, Arman Cohan, Nazli Goharian

In this work, we present a search system called SLEDGE, which utilizes SciBERT to effectively re-rank articles.

Fact or Fiction: Verifying Scientific Claims

2 code implementations EMNLP 2020 David Wadden, Shanchuan Lin, Kyle Lo, Lucy Lu Wang, Madeleine van Zuylen, Arman Cohan, Hannaneh Hajishirzi

We introduce scientific claim verification, a new task to select abstracts from the research literature containing evidence that SUPPORTS or REFUTES a given scientific claim, and to identify rationales justifying each decision.

Claim Verification Domain Adaptation +1

SPECTER: Document-level Representation Learning using Citation-informed Transformers

5 code implementations ACL 2020 Arman Cohan, Sergey Feldman, Iz Beltagy, Doug Downey, Daniel S. Weld

We propose SPECTER, a new method to generate document-level embedding of scientific documents based on pretraining a Transformer language model on a powerful signal of document-level relatedness: the citation graph.

Citation Prediction Document Classification +4

Longformer: The Long-Document Transformer

22 code implementations10 Apr 2020 Iz Beltagy, Matthew E. Peters, Arman Cohan

To address this limitation, we introduce the Longformer with an attention mechanism that scales linearly with sequence length, making it easy to process documents of thousands of tokens or longer.

Language Modelling Question Answering +1

Ranking Significant Discrepancies in Clinical Reports

no code implementations18 Jan 2020 Sean MacAvaney, Arman Cohan, Nazli Goharian, Ross Filice

This allows medical practitioners to easily identify and learn from the reports in which their interpretation most substantially differed from that of the attending physician (who finalized the report).

SUPP.AI: Finding Evidence for Supplement-Drug Interactions

1 code implementation ACL 2020 Lucy Lu Wang, Oyvind Tafjord, Arman Cohan, Sarthak Jain, Sam Skjonsberg, Carissa Schoenick, Nick Botner, Waleed Ammar

We fine-tune the contextualized word representations of the RoBERTa language model using labeled DDI data, and apply the fine-tuned model to identify supplement interactions.

General Classification Language Modelling

Pretrained Language Models for Sequential Sentence Classification

1 code implementation IJCNLP 2019 Arman Cohan, Iz Beltagy, Daniel King, Bhavana Dalvi, Daniel S. Weld

As a step toward better document-level understanding, we explore classification of a sequence of sentences into their corresponding categories, a task that requires understanding sentences in context of the document.

Classification General Classification +2

Ontology-Aware Clinical Abstractive Summarization

no code implementations14 May 2019 Sean MacAvaney, Sajad Sotudeh, Arman Cohan, Nazli Goharian, Ish Talati, Ross W. Filice

Automatically generating accurate summaries from clinical reports could save a clinician's time, improve summary coverage, and reduce errors.

Abstractive Text Summarization

Structural Scaffolds for Citation Intent Classification in Scientific Publications

1 code implementation NAACL 2019 Arman Cohan, Waleed Ammar, Madeleine van Zuylen, Field Cady

Identifying the intent of a citation in scientific papers (e. g., background information, use of methods, comparing results) is critical for machine reading of individual publications and automated analysis of the scientific literature.

Citation Intent Classification Classification +5

SciBERT: A Pretrained Language Model for Scientific Text

5 code implementations IJCNLP 2019 Iz Beltagy, Kyle Lo, Arman Cohan

Obtaining large-scale annotated data for NLP tasks in the scientific domain is challenging and expensive.

 Ranked #1 on Sentence Classification on Paper Field (using extra training data)

Citation Intent Classification Dependency Parsing +7

Depression and Self-Harm Risk Assessment in Online Forums

no code implementations EMNLP 2017 Andrew Yates, Arman Cohan, Nazli Goharian

We propose methods for identifying posts in support communities that may indicate a risk of self-harm, and demonstrate that our approach outperforms strong previously proposed methods for identifying such posts.

Identifying Harm Events in Clinical Care through Medical Narratives

no code implementations15 Aug 2017 Arman Cohan, Allan Fong, Raj Ratwani, Nazli Goharian

Preventable medical errors are estimated to be among the leading causes of injury and death in the United States.

Scientific document summarization via citation contextualization and scientific discourse

no code implementations12 Jun 2017 Arman Cohan, Nazli Goharian

We present a framework for scientific summarization which takes advantage of the citations and the scientific discourse structure.

Document Summarization Scientific Document Summarization +1

Contextualizing Citations for Scientific Summarization using Word Embeddings and Domain Knowledge

no code implementations23 May 2017 Arman Cohan, Nazli Goharian

Citation texts are sometimes not very informative or in some cases inaccurate by themselves; they need the appropriate context from the referenced paper to reflect its exact contributions.

Word Embeddings

Scientific Article Summarization Using Citation-Context and Article's Discourse Structure

1 code implementation EMNLP 2015 Arman Cohan, Nazli Goharian

We propose a summarization approach for scientific articles which takes advantage of citation-context and the document discourse model.

A Neural Attention Model for Categorizing Patient Safety Events

no code implementations23 Feb 2017 Arman Cohan, Allan Fong, Nazli Goharian, Raj Ratwani

Medical errors are leading causes of death in the US and as such, prevention of these errors is paramount to promoting health care.

Triaging Content Severity in Online Mental Health Forums

no code implementations22 Feb 2017 Arman Cohan, Sydney Young, Andrew Yates, Nazli Goharian

Our analysis on the interaction of the moderators with the users further indicates that without an automatic way to identify critical content, it is indeed challenging for the moderators to provide timely response to the users in need.

Revisiting Summarization Evaluation for Scientific Articles

1 code implementation LREC 2016 Arman Cohan, Nazli Goharian

Finally, we propose an alternative metric for summarization evaluation which is based on the content relevance between a system generated summary and the corresponding human written summaries.

Text Summarization

Cannot find the paper you are looking for? You can Submit a new open access paper.