Search Results for author: Huiyi Chen

Found 3 papers, 1 papers with code

Causality-based Cross-Modal Representation Learning for Vision-and-Language Navigation

no code implementations6 Mar 2024 Liuyi Wang, Zongtao He, Ronghao Dang, Huiyi Chen, Chengju Liu, Qijun Chen

Vision-and-Language Navigation (VLN) has gained significant research interest in recent years due to its potential applications in real-world scenarios.

Representation Learning Vision and Language Navigation

How to Configure Good In-Context Sequence for Visual Question Answering

1 code implementation4 Dec 2023 Li Li, Jiawei Peng, Huiyi Chen, Chongyang Gao, Xu Yang

Inspired by the success of Large Language Models in dealing with new tasks via In-Context Learning (ICL) in NLP, researchers have also developed Large Vision-Language Models (LVLMs) with ICL capabilities.

In-Context Learning Question Answering +2

PASTS: Progress-Aware Spatio-Temporal Transformer Speaker For Vision-and-Language Navigation

no code implementations19 May 2023 Liuyi Wang, Chengju Liu, Zongtao He, Shu Li, Qingqing Yan, Huiyi Chen, Qijun Chen

The experimental results demonstrate that PASTS outperforms all existing speaker models and successfully improves the performance of previous VLN models, achieving state-of-the-art performance on the standard Room-to-Room (R2R) dataset.

Data Augmentation Vision and Language Navigation

Cannot find the paper you are looking for? You can Submit a new open access paper.