Search Results for author: Zongxia Li

Found 6 papers, 5 papers with code

PANDA (Pedantic ANswer-correctness Determination and Adjudication):Improving Automatic Evaluation for Question Answering and Text Generation

1 code implementation17 Feb 2024 Zongxia Li, Ishani Mondal, Yijun Liang, Huy Nghiem, Jordan Lee Boyd-Graber

Question answering (QA) can only make progress if we know if an answer is correct, but for many of the most challenging and interesting QA examples, current answer correctness (AC) metrics do not align with human judgments, particularly verbose, free form answers from large language models (LLM).

Question Answering Text Generation

Improving the TENOR of Labeling: Re-evaluating Topic Models for Content Analysis

1 code implementation29 Jan 2024 Zongxia Li, Andrew Mao, Daniel Stephens, Pranav Goel, Emily Walpole, Alden Dima, Juan Fung, Jordan Boyd-Graber

Topic models are a popular tool for understanding text collections, but their evaluation has been a point of contention.

Topic Models

CFMatch: Aligning Automated Answer Equivalence Evaluation with Expert Judgments For Open-Domain Question Answering

no code implementations24 Jan 2024 Zongxia Li, Ishani Mondal, Yijun Liang, Huy Nghiem, Jordan Boyd-Graber

Question answering (QA) can only make progress if we know if an answer is correct, but for many of the most challenging and interesting QA examples, current evaluation metrics to determine answer equivalence (AE) often do not align with human judgments, particularly more verbose, free-form answers from large language models (LLM).

Open-Domain Question Answering

Towards Understanding In-Context Learning with Contrastive Demonstrations and Saliency Maps

1 code implementation11 Jul 2023 Fuxiao Liu, Paiheng Xu, Zongxia Li, Yue Feng

We investigate the role of various demonstration components in the in-context learning (ICL) performance of large language models (LLMs).

In-Context Learning Sentiment Analysis

SODAPOP: Open-Ended Discovery of Social Biases in Social Commonsense Reasoning Models

1 code implementation13 Oct 2022 Haozhe An, Zongxia Li, Jieyu Zhao, Rachel Rudinger

A common limitation of diagnostic tests for detecting social biases in NLP models is that they may only detect stereotypic associations that are pre-specified by the designer of the test.

Language Modelling Question Answering

Cannot find the paper you are looking for? You can Submit a new open access paper.