Search Results for author: Cunxiang Wang

Found 19 papers, 12 papers with code

NovelQA: A Benchmark for Long-Range Novel Question Answering

1 code implementation18 Mar 2024 Cunxiang Wang, Ruoxi Ning, Boqi Pan, Tonghui Wu, Qipeng Guo, Cheng Deng, Guangsheng Bao, Qian Wang, Yue Zhang

The rapid advancement of Large Language Models (LLMs) has introduced a new frontier in natural language processing, particularly in understanding and processing long-context information.

Question Answering

Knowledge Conflicts for LLMs: A Survey

no code implementations13 Mar 2024 Rongwu Xu, Zehan Qi, Cunxiang Wang, Hongru Wang, Yue Zhang, Wei Xu

This survey provides an in-depth analysis of knowledge conflicts for large language models (LLMs), highlighting the complex challenges they encounter when blending contextual and parametric knowledge.

Misinformation

LLMs with Chain-of-Thought Are Non-Causal Reasoners

1 code implementation25 Feb 2024 Guangsheng Bao, Hongbo Zhang, Linyi Yang, Cunxiang Wang, Yue Zhang

We further examine the factors influencing the causal structure of the implied SCM, revealing that in-context learning, supervised fine-tuning, and reinforcement learning on human feedback significantly impact the causal relations.

In-Context Learning

Self-DC: When to retrieve and When to generate? Self Divide-and-Conquer for Compositional Unknown Questions

no code implementations21 Feb 2024 Hongru Wang, Boyang Xue, Baohang Zhou, Tianhua Zhang, Cunxiang Wang, Guanhua Chen, Huimin Wang, Kam-Fai Wong

Retrieve-then-read and generate-then-read are two typical solutions to handle unknown and known questions in open-domain question-answering, while the former retrieves necessary external knowledge and the later prompt the large language models to generate internal known knowledge encoded in the parameters.

Binary Classification Open-Domain Question Answering +1

SQL-CRAFT: Text-to-SQL through Interactive Refinement and Enhanced Reasoning

no code implementations20 Feb 2024 Hanchen Xia, Feng Jiang, Naihao Deng, Cunxiang Wang, Guojiang Zhao, Rada Mihalcea, Yue Zhang

Modern LLMs have become increasingly powerful, but they are still facing challenges in specialized tasks such as Text-to-SQL.

Text-To-SQL

TRAMS: Training-free Memory Selection for Long-range Language Modeling

1 code implementation24 Oct 2023 Haofei Yu, Cunxiang Wang, Yue Zhang, Wei Bi

The Transformer architecture is crucial for numerous AI models, but it still faces challenges in long-range language modeling.

Language Modelling

A Survey on Evaluation of Large Language Models

1 code implementation6 Jul 2023 Yupeng Chang, Xu Wang, Jindong Wang, Yuan Wu, Linyi Yang, Kaijie Zhu, Hao Chen, Xiaoyuan Yi, Cunxiang Wang, Yidong Wang, Wei Ye, Yue Zhang, Yi Chang, Philip S. Yu, Qiang Yang, Xing Xie

Large language models (LLMs) are gaining increasing popularity in both academia and industry, owing to their unprecedented performance in various applications.

Ethics

PandaLM: An Automatic Evaluation Benchmark for LLM Instruction Tuning Optimization

2 code implementations8 Jun 2023 Yidong Wang, Zhuohao Yu, Zhengran Zeng, Linyi Yang, Cunxiang Wang, Hao Chen, Chaoya Jiang, Rui Xie, Jindong Wang, Xing Xie, Wei Ye, Shikun Zhang, Yue Zhang

To ensure the reliability of PandaLM, we collect a diverse human-annotated test dataset, where all contexts are generated by humans and labels are aligned with human preferences.

Language Modelling Large Language Model

RFiD: Towards Rational Fusion-in-Decoder for Open-Domain Question Answering

1 code implementation26 May 2023 Cunxiang Wang, Haofei Yu, Yue Zhang

Open-Domain Question Answering (ODQA) systems necessitate a reader model capable of generating answers by simultaneously referring to multiple passages.

Natural Questions Open-Domain Question Answering +1

Exploiting Abstract Meaning Representation for Open-Domain Question Answering

1 code implementation26 May 2023 Cunxiang Wang, Zhikun Xu, Qipeng Guo, Xiangkun Hu, Xuefeng Bai, Zheng Zhang, Yue Zhang

The Open-Domain Question Answering (ODQA) task involves retrieving and subsequently generating answers from fine-grained relevant passages within a database.

Natural Questions Open-Domain Question Answering +1

Evaluating Open-QA Evaluation

1 code implementation NeurIPS 2023 Cunxiang Wang, Sirui Cheng, Qipeng Guo, Yuanhao Yue, Bowen Ding, Zhikun Xu, Yidong Wang, Xiangkun Hu, Zheng Zhang, Yue Zhang

This study focuses on the evaluation of the Open Question Answering (Open-QA) task, which can directly estimate the factuality of large language models (LLMs).

Question Answering

Knowledgeable Salient Span Mask for Enhancing Language Models as Knowledge Base

no code implementations17 Apr 2022 Cunxiang Wang, Fuli Luo, Yanyang Li, Runxin Xu, Fei Huang, Yue Zhang

Pre-trained language models (PLMs) like BERT have made significant progress in various downstream NLP tasks.

Self-Supervised Learning

Exploring Generalization Ability of Pretrained Language Models on Arithmetic and Logical Reasoning

no code implementations15 Aug 2021 Cunxiang Wang, Boyuan Zheng, Yuchen Niu, Yue Zhang

To quantitatively and intuitively explore the generalization ability of pre-trained language models (PLMs), we have designed several tasks of arithmetic and logical reasoning.

Logical Reasoning

Can Generative Pre-trained Language Models Serve as Knowledge Bases for Closed-book QA?

1 code implementation ACL 2021 Cunxiang Wang, Pai Liu, Yue Zhang

Recent work has investigated the interesting question using pre-trained language models (PLMs) as knowledge bases for answering open questions.

Question Answering

Commonsense Knowledge Graph Reasoning by Selection or Generation? Why?

no code implementations13 Aug 2020 Cunxiang Wang, Jinhang Wu, Luxin Liu, Yue Zhang

We theoretically and empirically compare the two methods, finding the selection method is more suitable than the generation method in CKGR.

Knowledge Graph Embedding

SemEval-2020 Task 4: Commonsense Validation and Explanation

2 code implementations SEMEVAL 2020 Cunxiang Wang, Shuailong Liang, Yili Jin, Yilong Wang, Xiaodan Zhu, Yue Zhang

In this paper, we present SemEval-2020 Task 4, Commonsense Validation and Explanation (ComVE), which includes three subtasks, aiming to evaluate whether a system can distinguish a natural language statement that makes sense to humans from one that does not, and provide the reasons.

Domain Representation for Knowledge Graph Embedding

no code implementations26 Mar 2019 Cunxiang Wang, Feiliang Ren, Zhichao Lin, Chenxv Zhao, Tian Xie, Yue Zhang

Embedding entities and relations into a continuous multi-dimensional vector space have become the dominant method for knowledge graph embedding in representation learning.

Knowledge Graph Embedding Link Prediction +1

Cannot find the paper you are looking for? You can Submit a new open access paper.