Search Results for author: Chengzu Li

Found 8 papers, 7 papers with code

A Call for New Recipes to Enhance Spatial Reasoning in MLLMs

no code implementations21 Apr 2025 Huanyu Zhang, Chengzu Li, Wenshan Wu, Shaoguang Mao, Yan Xia, Ivan Vulić, Zhang Zhang, Liang Wang, Tieniu Tan, Furu Wei

Multimodal Large Language Models (MLLMs) have demonstrated impressive performance in general vision-language tasks.

Spatial Reasoning

Imagine while Reasoning in Space: Multimodal Visualization-of-Thought

1 code implementation13 Jan 2025 Chengzu Li, Wenshan Wu, Huanyu Zhang, Yan Xia, Shaoguang Mao, Li Dong, Ivan Vulić, Furu Wei

Ultimately, MVoT establishes new possibilities for complex reasoning tasks where visual thinking can effectively complement verbal reasoning.

Spatial Reasoning

TopViewRS: Vision-Language Models as Top-View Spatial Reasoners

1 code implementation4 Jun 2024 Chengzu Li, Caiqi Zhang, Han Zhou, Nigel Collier, Anna Korhonen, Ivan Vulić

In this work, we thus study their capability to understand and reason over spatial relations from the top view.

Multiple-choice Spatial Reasoning

Semantic Map-based Generation of Navigation Instructions

1 code implementation28 Mar 2024 Chengzu Li, Chao Zhang, Simone Teufel, Rama Sanand Doddipatla, Svetlana Stoyanchev

In this paper, we propose a new approach to navigation instruction generation by framing the problem as an image captioning task using semantic maps as visual input.

Image Captioning

On Task Performance and Model Calibration with Supervised and Self-Ensembled In-Context Learning

1 code implementation21 Dec 2023 Chengzu Li, Han Zhou, Goran Glavaš, Anna Korhonen, Ivan Vulić

Following the standard supervised fine-tuning (SFT) paradigm, in-context learning (ICL) has become an efficient approach propelled by the recent advancements in large language models (LLMs), yielding promising performance across various tasks in few-shot data setups.

In-Context Learning

Generating Data for Symbolic Language with Large Language Models

1 code implementation23 May 2023 Jiacheng Ye, Chengzu Li, Lingpeng Kong, Tao Yu

However, such an approach has primarily been applied to natural language tasks and has not yet been explored for symbolic language tasks with complex structured outputs (e. g., semantic parsing and code generation).

Code Generation Semantic Parsing

Binding Language Models in Symbolic Languages

4 code implementations6 Oct 2022 Zhoujun Cheng, Tianbao Xie, Peng Shi, Chengzu Li, Rahul Nadkarni, Yushi Hu, Caiming Xiong, Dragomir Radev, Mari Ostendorf, Luke Zettlemoyer, Noah A. Smith, Tao Yu

We propose Binder, a training-free neural-symbolic framework that maps the task input to a program, which (1) allows binding a unified API of language model (LM) functionalities to a programming language (e. g., SQL, Python) to extend its grammar coverage and thus tackle more diverse questions, (2) adopts an LM as both the program parser and the underlying model called by the API during execution, and (3) requires only a few in-context exemplar annotations.

Language Modelling Semantic Parsing +1

Cannot find the paper you are looking for? You can Submit a new open access paper.