Search Results for author: Xingxuan Li

Found 14 papers, 10 papers with code

Chain of Ideas: Revolutionizing Research Via Novel Idea Development with LLM Agents

1 code implementation17 Oct 2024 Long Li, Weiwen Xu, Jiayan Guo, Ruochen Zhao, Xingxuan Li, Yuqian Yuan, Boqiang Zhang, Yuming Jiang, Yifei Xin, Ronghao Dang, Deli Zhao, Yu Rong, Tian Feng, Lidong Bing

Moreover, our CoI agent is budget-friendly, with a minimum cost of \$0. 50 to generate a candidate idea and its corresponding experimental design.

Experimental Design

Can We Further Elicit Reasoning in LLMs? Critic-Guided Planning with Retrieval-Augmentation for Solving Challenging Tasks

no code implementations2 Oct 2024 Xingxuan Li, Weiwen Xu, Ruochen Zhao, Fangkai Jiao, Shafiq Joty, Lidong Bing

We validate CR-Planner on challenging domain-knowledge-intensive and reasoning-heavy tasks, including competitive programming, theorem-driven math reasoning, and complex domain retrieval problems.

Math Navigate +2

How Much are Large Language Models Contaminated? A Comprehensive Survey and the LLMSanitize Library

1 code implementation31 Mar 2024 Mathieu Ravaut, Bosheng Ding, Fangkai Jiao, Hailin Chen, Xingxuan Li, Ruochen Zhao, Chengwei Qin, Caiming Xiong, Shafiq Joty

With the rise of Large Language Models (LLMs) in recent years, abundant new opportunities are emerging, but also new challenges, among which contamination is quickly becoming critical.

Question Answering

ParaICL: Towards Robust Parallel In-Context Learning

no code implementations31 Mar 2024 Xingxuan Li, Xuan-Phi Nguyen, Shafiq Joty, Lidong Bing

However, our preliminary experiments indicate that the effectiveness of ICL is limited by the length of the input context.

In-Context Learning Semantic Similarity +1

SeaLLMs -- Large Language Models for Southeast Asia

1 code implementation1 Dec 2023 Xuan-Phi Nguyen, Wenxuan Zhang, Xin Li, Mahani Aljunied, Zhiqiang Hu, Chenhui Shen, Yew Ken Chia, Xingxuan Li, Jianyu Wang, Qingyu Tan, Liying Cheng, Guanzheng Chen, Yue Deng, Sen yang, Chaoqun Liu, Hang Zhang, Lidong Bing

Despite the remarkable achievements of large language models (LLMs) in various tasks, there remains a linguistic bias that favors high-resource languages, such as English, often at the expense of low-resource and regional languages.

Instruction Following

Is GPT-4 a Good Data Analyst?

1 code implementation24 May 2023 Liying Cheng, Xingxuan Li, Lidong Bing

As large language models (LLMs) have demonstrated their powerful capabilities in plenty of domains and tasks, including context understanding, code generation, language generation, data storytelling, etc., many data analysts may raise concerns if their jobs will be replaced by artificial intelligence (AI).

Code Generation Text Generation

Verify-and-Edit: A Knowledge-Enhanced Chain-of-Thought Framework

1 code implementation5 May 2023 Ruochen Zhao, Xingxuan Li, Shafiq Joty, Chengwei Qin, Lidong Bing

As large language models (LLMs) have become the norm in NLP, demonstrating good performance in generation and reasoning tasks, one of its most fatal disadvantages is the lack of factual correctness.

Open-Domain Question Answering

Retrieving Multimodal Information for Augmented Generation: A Survey

no code implementations20 Mar 2023 Ruochen Zhao, Hailin Chen, Weishi Wang, Fangkai Jiao, Xuan Long Do, Chengwei Qin, Bosheng Ding, Xiaobao Guo, Minzhi Li, Xingxuan Li, Shafiq Joty

As Large Language Models (LLMs) become popular, there emerged an important trend of using multimodality to augment the LLMs' generation ability, which enables LLMs to better interact with the world.

Retrieval Survey

Evaluating Psychological Safety of Large Language Models

no code implementations20 Dec 2022 Xingxuan Li, Yutong Li, Lin Qiu, Shafiq Joty, Lidong Bing

Despite being instruction fine-tuned with safety metrics to reduce toxicity, InstructGPT, GPT-3. 5, and GPT-4 still showed dark personality patterns; these models scored higher than self-supervised GPT-3 on the Machiavellianism and narcissism traits on SD-3.

Towards Robust Low-Resource Fine-Tuning with Multi-View Compressed Representations

1 code implementation16 Nov 2022 Linlin Liu, Xingxuan Li, Megh Thakkar, Xin Li, Shafiq Joty, Luo Si, Lidong Bing

Due to the huge amount of parameters, fine-tuning of pretrained language models (PLMs) is prone to overfitting in the low resource scenarios.

YEDDA: A Lightweight Collaborative Text Span Annotation Tool

1 code implementation ACL 2018 Jie Yang, Yue Zhang, Linwei Li, Xingxuan Li

And the annotation time can be further compressed by 16. 47\% through intelligent recommendation.

text annotation

Cannot find the paper you are looking for? You can Submit a new open access paper.