Search Results for author: Naihao Deng

Found 21 papers, 7 papers with code

In-the-Wild Video Question Answering

no code implementations COLING 2022 Santiago Castro, Naihao Deng, Pingxuan Huang, Mihai Burzo, Rada Mihalcea

Existing video understanding datasets mostly focus on human interactions, with little attention being paid to the “in the wild” settings, where the videos are recorded outdoors.

Evidence Selection Question Answering +2

Analyzing the Effects of Annotator Gender across NLP Tasks

1 code implementation NLPerspectives (LREC) 2022 Laura Biester, Vanita Sharma, Ashkan Kazemi, Naihao Deng, Steven Wilson, Rada Mihalcea

Recent studies have shown that for subjective annotation tasks, the demographics, lived experiences, and identity of annotators can have a large impact on how items are labeled.

Natural Language Inference

Are Human Interactions Replicable by Generative Agents? A Case Study on Pronoun Usage in Hierarchical Interactions

no code implementations25 Jan 2025 Naihao Deng, Rada Mihalcea

As Large Language Models (LLMs) advance in their capabilities, researchers have increasingly employed them for social simulation.

Decision Making

Rethinking Table Instruction Tuning

no code implementations24 Jan 2025 Naihao Deng, Rada Mihalcea

Recent advances in table understanding have focused on instruction-tuning large language models (LLMs) for table-related tasks.

Domain Generalization

Table as Thought: Exploring Structured Thoughts in LLM Reasoning

no code implementations4 Jan 2025 Zhenjie Sun, Naihao Deng, Haofei Yu, Jiaxuan You

Large language models' reasoning abilities benefit from methods that organize their thought processes, such as chain-of-thought prompting, which employs a sequential structure to guide the reasoning process step-by-step.

Mathematical Reasoning

Chumor 2.0: Towards Benchmarking Chinese Humor Understanding

no code implementations23 Dec 2024 Ruiqi He, Yushu He, Longju Bai, Jiarui Liu, Zhenjie Sun, Zenghao Tang, He Wang, Hanchen Xia, Rada Mihalcea, Naihao Deng

Existing humor datasets and evaluations predominantly focus on English, leaving limited resources for culturally nuanced humor in non-English languages like Chinese.

Benchmarking

Chumor 1.0: A Truly Funny and Challenging Chinese Humor Understanding Dataset from Ruo Zhi Ba

no code implementations18 Jun 2024 Ruiqi He, Yushu He, Longju Bai, Jiarui Liu, Zhenjie Sun, Zenghao Tang, He Wang, Hanchen Xia, Naihao Deng

Existing humor datasets and evaluations predominantly focus on English, lacking resources for culturally nuanced humor in non-English languages like Chinese.

Tables as Texts or Images: Evaluating the Table Reasoning Ability of LLMs and MLLMs

no code implementations19 Feb 2024 Naihao Deng, Zhenjie Sun, Ruiqi He, Aman Sikka, Yulong Chen, Lin Ma, Yue Zhang, Rada Mihalcea

In this paper, we investigate the effectiveness of various LLMs in interpreting tabular data through different prompting strategies and data formats.

Fact Checking Question Answering

Task-Adaptive Tokenization: Enhancing Long-Form Text Generation Efficacy in Mental Health and Beyond

1 code implementation9 Oct 2023 Siyang Liu, Naihao Deng, Sahand Sabour, Yilin Jia, Minlie Huang, Rada Mihalcea

We propose task-adaptive tokenization as a way to adapt the generation pipeline to the specifics of a downstream task and enhance long-form generation in mental health.

Form Question Answering +1

EASE: An Easily-Customized Annotation System Powered by Efficiency Enhancement Mechanisms

no code implementations23 May 2023 Naihao Deng, YiKai Liu, Mingye Chen, Winston Wu, Siyang Liu, Yulong Chen, Yue Zhang, Rada Mihalcea

Our results show that our system can meet the diverse needs of NLP researchers and significantly accelerate the annotation process.

Active Learning

Query Rewriting for Effective Misinformation Discovery

no code implementations14 Oct 2022 Ashkan Kazemi, Artem Abzaliev, Naihao Deng, Rui Hou, Scott A. Hale, Verónica Pérez-Rosas, Rada Mihalcea

We propose a novel system to help fact-checkers formulate search queries for known misinformation claims and effectively search across multiple social media platforms.

Misinformation reinforcement-learning +2

WildQA: In-the-Wild Video Question Answering

no code implementations14 Sep 2022 Santiago Castro, Naihao Deng, Pingxuan Huang, Mihai Burzo, Rada Mihalcea

Existing video understanding datasets mostly focus on human interactions, with little attention being paid to the "in the wild" settings, where the videos are recorded outdoors.

Evidence Selection Question Answering +2

Recent Advances in Text-to-SQL: A Survey of What We Have and What We Expect

1 code implementation COLING 2022 Naihao Deng, Yulong Chen, Yue Zhang

Text-to-SQL has attracted attention from both the natural language processing and database communities because of its ability to convert the semantics in natural language into SQL queries and its practical application in building natural language interfaces to database systems.

Survey Text-To-SQL

DialogSum Challenge: Results of the Dialogue Summarization Shared Task

1 code implementation8 Aug 2022 Yulong Chen, Naihao Deng, Yang Liu, Yue Zhang

We report the results of DialogSum Challenge, the shared task on summarizing real-life scenario dialogues at INLG 2022.

The Cross-lingual Conversation Summarization Challenge

2 code implementations1 May 2022 Yulong Chen, Ming Zhong, Xuefeng Bai, Naihao Deng, Jing Li, Xianchao Zhu, Yue Zhang

We propose the shared task of cross-lingual conversation summarization, \emph{ConvSumX Challenge}, opening new avenues for researchers to investigate solutions that integrate conversation summarization and machine translation.

Abstractive Dialogue Summarization Conversation Summarization +4

Prefix-to-SQL: Text-to-SQL Generation from Incomplete User Questions

no code implementations15 Sep 2021 Naihao Deng, Shuaichen Chang, Peng Shi, Tao Yu, Rui Zhang

Existing text-to-SQL research only considers complete questions as the input, but lay-users might strive to formulate a complete question.

Text-To-SQL

Cannot find the paper you are looking for? You can Submit a new open access paper.