Search Results for author: Jiabang He

Found 6 papers, 5 papers with code

Mitigating Fine-Grained Hallucination by Fine-Tuning Large Vision-Language Models with Caption Rewrites

1 code implementation4 Dec 2023 Lei Wang, Jiabang He, Shenshen Li, Ning Liu, Ee-Peng Lim

The fine-grained object attributes and behaviors non-existent in the image may still be generated but not measured by the current evaluation methods.

Hallucination Hallucination Evaluation +2

MoCoSA: Momentum Contrast for Knowledge Graph Completion with Structure-Augmented Pre-trained Language Models

no code implementations16 Aug 2023 Jiabang He, Liu Jia, Lei Wang, Xiyao Li, Xing Xu

However, they struggle with semantically rich real-world entities due to limited structural information and fail to generalize to unseen entities.

Entity Embeddings Link Prediction

Do-GOOD: Towards Distribution Shift Evaluation for Pre-Trained Visual Document Understanding Models

1 code implementation5 Jun 2023 Jiabang He, Yi Hu, Lei Wang, Xing Xu, Ning Liu, Hui Liu, Heng Tao Shen

Results from the experiments demonstrate that there is a significant performance gap between the in-distribution (ID) and OOD settings for document images, and that fine-grained analysis of distribution shifts can reveal the brittle nature of existing pre-trained VDU models and OOD generalization algorithms.

document understanding Question Answering

ICL-D3IE: In-Context Learning with Diverse Demonstrations Updating for Document Information Extraction

1 code implementation ICCV 2023 Jiabang He, Lei Wang, Yi Hu, Ning Liu, Hui Liu, Xing Xu, Heng Tao Shen

To this end, we propose a simple but effective in-context learning framework called ICL-D3IE, which enables LLMs to perform DIE with different types of demonstration examples.

Document AI In-Context Learning

Alignment-Enriched Tuning for Patch-Level Pre-trained Document Image Models

1 code implementation27 Nov 2022 Lei Wang, Jiabang He, Xing Xu, Ning Liu, Hui Liu

In this paper, we propose a new model architecture with alignment-enriched tuning (dubbed AETNet) upon pre-trained document image models, to adapt downstream tasks with the joint task-specific supervised and alignment-aware contrastive objective.

Cannot find the paper you are looking for? You can Submit a new open access paper.