Search Results for author: Linyong Nan

Found 17 papers, 14 papers with code

DocMath-Eval: Evaluating Numerical Reasoning Capabilities of LLMs in Understanding Long Documents with Tabular Data

no code implementations16 Nov 2023 Yilun Zhao, Yitao Long, Hongjun Liu, Linyong Nan, Lyuhao Chen, Ryo Kamoi, Yixin Liu, Xiangru Tang, Rui Zhang, Arman Cohan

This paper introduces DocMath-Eval, a comprehensive benchmark specifically designed to evaluate the numerical reasoning and problem-solving capabilities of LLMs in the context of understanding and analyzing financial documents containing both text and tables.

Math

On Evaluating the Integration of Reasoning and Action in LLM Agents with Database Question Answering

no code implementations16 Nov 2023 Linyong Nan, Ellen Zhang, Weijin Zou, Yilun Zhao, Wenfei Zhou, Arman Cohan

A key discovery is the identification of two primary bottlenecks hindering effective interaction: the capacity for planning and the ability to generate multiple SQL queries.

Question Answering Retrieval

RobuT: A Systematic Study of Table QA Robustness Against Human-Annotated Adversarial Perturbations

1 code implementation25 Jun 2023 Yilun Zhao, Chen Zhao, Linyong Nan, Zhenting Qi, Wenlin Zhang, Xiangru Tang, Boyu Mi, Dragomir Radev

Despite significant progress having been made in question answering on tabular data (Table QA), it's unclear whether, and to what extent existing Table QA models are robust to task-specific perturbations, e. g., replacing key question entities or shuffling table columns.

Few-Shot Learning Question Answering

Investigating Table-to-Text Generation Capabilities of LLMs in Real-World Information Seeking Scenarios

2 code implementations24 May 2023 Yilun Zhao, Haowei Zhang, Shengyun Si, Linyong Nan, Xiangru Tang, Arman Cohan

These include the LogicNLG and our newly-constructed LoTNLG datasets for data insight generation, along with the FeTaQA and our newly-constructed F2WTQ datasets for query-based generation.

Table-to-Text Generation

QTSumm: Query-Focused Summarization over Tabular Data

2 code implementations23 May 2023 Yilun Zhao, Zhenting Qi, Linyong Nan, Boyu Mi, Yixin Liu, Weijin Zou, Simeng Han, Ruizhe Chen, Xiangru Tang, Yumo Xu, Dragomir Radev, Arman Cohan

Motivated by this, we define a new query-focused table summarization task, where text generation models have to perform human-like reasoning and analysis over the given table to generate a tailored summary.

Query-focused Summarization Table-to-Text Generation

Enhancing Few-shot Text-to-SQL Capabilities of Large Language Models: A Study on Prompt Design Strategies

no code implementations21 May 2023 Linyong Nan, Yilun Zhao, Weijin Zou, Narutatsu Ri, Jaesung Tae, Ellen Zhang, Arman Cohan, Dragomir Radev

In-context learning (ICL) has emerged as a new approach to various natural language processing tasks, utilizing large language models (LLMs) to make predictions based on context that has been supplemented with a few examples or task-specific instructions.

In-Context Learning Question Answering +1

ReasTAP: Injecting Table Reasoning Skills During Pre-training via Synthetic Reasoning Examples

1 code implementation22 Oct 2022 Yilun Zhao, Linyong Nan, Zhenting Qi, Rui Zhang, Dragomir Radev

Reasoning over tabular data requires both table structure understanding and a broad set of table reasoning skills.

Ranked #3 on Semantic Parsing on WikiSQL (Denotation accuracy (test) metric)

Fact Verification Question Answering +3

Leveraging Locality in Abstractive Text Summarization

1 code implementation25 May 2022 Yixin Liu, Ansong Ni, Linyong Nan, Budhaditya Deb, Chenguang Zhu, Ahmed H. Awadallah, Dragomir Radev

Our experimental results show that our model has a better performance compared with strong baselines with efficient attention modules, and our analysis provides further insights into our locality-aware modeling strategy.

Abstractive Text Summarization Text Generation

FeTaQA: Free-form Table Question Answering

1 code implementation1 Apr 2021 Linyong Nan, Chiachun Hsieh, Ziming Mao, Xi Victoria Lin, Neha Verma, Rui Zhang, Wojciech Kryściński, Nick Schoelkopf, Riley Kong, Xiangru Tang, Murori Mutuma, Ben Rosand, Isabel Trindade, Renusree Bandaru, Jacob Cunningham, Caiming Xiong, Dragomir Radev

Existing table question answering datasets contain abundant factual questions that primarily evaluate the query and schema comprehension capability of a system, but they fail to include questions that require complex reasoning and integration of information due to the constraint of the associated short-form answers.

Question Answering Retrieval +2

Detecting Urgency Status of Crisis Tweets: A Transfer Learning Approach for Low Resource Languages

1 code implementation COLING 2020 Efsun Sarioglu Kayi, Linyong Nan, Bohan Qu, Mona Diab, Kathleen McKeown

We adopt cross-lingual embeddings constructed using different methods to extract features of the tweets, including a few state-of-the-art contextual embeddings such as BERT, RoBERTa and XLM-R. We train classifiers of different architectures on the extracted features.

Transfer Learning XLM-R

Accurate Protein Structure Prediction by Embeddings and Deep Learning Representations

3 code implementations9 Nov 2019 Iddo Drori, Darshan Thaker, Arjun Srivatsa, Daniel Jeong, Yueqi Wang, Linyong Nan, Fan Wu, Dimitri Leggas, Jinhao Lei, Weiyi Lu, Weilong Fu, Yuan Gao, Sashank Karri, Anand Kannan, Antonio Moretti, Mohammed AlQuraishi, Chen Keasar, Itsik Pe'er

Our dataset consists of amino acid sequences, Q8 secondary structures, position specific scoring matrices, multiple sequence alignment co-evolutionary features, backbone atom distance matrices, torsion angles, and 3D coordinates.

Multiple Sequence Alignment Protein Structure Prediction

Cannot find the paper you are looking for? You can Submit a new open access paper.