Search Results for author: Fengbin Zhu

Found 17 papers, 5 papers with code

FinTMMBench: Benchmarking Temporal-Aware Multi-Modal RAG in Finance

no code implementations7 Mar 2025 Fengbin Zhu, Junfeng Li, Liangming Pan, Wenjie Wang, Fuli Feng, Chao Wang, Huanbo Luan, Tat-Seng Chua

Finance decision-making often relies on in-depth data analysis across various data sources, including financial tables, news articles, stock prices, etc.

Benchmarking Event Detection +4

Large Language Models for Recommendation with Deliberative User Preference Alignment

no code implementations4 Feb 2025 Yi Fang, Wenjie Wang, Yang Zhang, Fengbin Zhu, Qifan Wang, Fuli Feng, Xiangnan He

We then introduce the Deliberative User Preference Alignment framework, designed to enhance reasoning capabilities by utilizing verbalized user feedback in a step-wise manner to tackle this task.

Leveraging Memory Retrieval to Enhance LLM-based Generative Recommendation

no code implementations23 Dec 2024 Chengbing Wang, Yang Zhang, Fengbin Zhu, Jizhi Zhang, Tianhao Shi, Fuli Feng

Leveraging Large Language Models (LLMs) to harness user-item interaction histories for item generation has emerged as a promising paradigm in generative recommendation.

Retrieval

Cooperative SQL Generation for Segmented Databases By Using Multi-functional LLM Agents

no code implementations8 Dec 2024 Zhiguang Wu, Fengbin Zhu, Xuequn Shang, Yupei Zhang, Pan Zhou

In the first stage, agents analyze their respective schema and communicate with each other to collect the schema information relevant to the question.

Language Modeling Language Modelling +2

Effective and Efficient Adversarial Detection for Vision-Language Models via A Single Vector

1 code implementation30 Oct 2024 Youcheng Huang, Fengbin Zhu, Jingkun Tang, Pan Zhou, Wenqiang Lei, Jiancheng Lv, Tat-Seng Chua

With the new RADAR dataset, we further develop a novel and effective iN-time Embedding-based AdveRSarial Image DEtection (NEARSIDE) method, which exploits a single vector that distilled from the hidden states of VLMs, which we call the attacking direction, to achieve the detection of adversarial images against benign ones in the input.

MMDocBench: Benchmarking Large Vision-Language Models for Fine-Grained Visual Document Understanding

no code implementations25 Oct 2024 Fengbin Zhu, Ziyang Liu, Xiang Yao Ng, Haohui Wu, Wenjie Wang, Fuli Feng, Chao Wang, Huanbo Luan, Tat Seng Chua

Large Vision-Language Models (LVLMs) have achieved remarkable performance in many vision-language tasks, yet their capabilities in fine-grained visual understanding remain insufficiently evaluated.

Benchmarking document understanding +1

Large Language Models Empowered Personalized Web Agents

no code implementations22 Oct 2024 Hongru Cai, Yongqi Li, Wenjie Wang, Fengbin Zhu, Xiaoyu Shen, Wenjie Li, Tat-Seng Chua

To overcome the limitation, we first formulate the task of LLM-empowered personalized Web agents, which integrate personalized data and user instructions to personalize instruction comprehension and action execution.

CrAM: Credibility-Aware Attention Modification in LLMs for Combating Misinformation in RAG

1 code implementation17 Jun 2024 Boyi Deng, Wenjie Wang, Fengbin Zhu, Qifan Wang, Fuli Feng

To address this issue, we explore the task of "credibility-aware RAG", in which LLMs automatically adjust the influence of retrieved documents based on their credibility scores to counteract misinformation.

Misinformation RAG +2

Think Twice Before Trusting: Self-Detection for Large Language Models through Comprehensive Answer Reflection

no code implementations15 Mar 2024 Moxin Li, Wenjie Wang, Fuli Feng, Fengbin Zhu, Qifan Wang, Tat-Seng Chua

Self-detection for Large Language Models (LLMs) seeks to evaluate the trustworthiness of the LLM's output by leveraging its own capabilities, thereby alleviating the issue of output hallucination.

Hallucination Language Modelling +1

TAT-LLM: A Specialized Language Model for Discrete Reasoning over Tabular and Textual Data

no code implementations24 Jan 2024 Fengbin Zhu, Ziyang Liu, Fuli Feng, Chao Wang, Moxin Li, Tat-Seng Chua

In this work, we address question answering (QA) over a hybrid of tabular and textual data that are very common content on the Web (e. g. SEC filings), where discrete reasoning capabilities are often required.

Language Modeling Language Modelling +1

Doc2SoarGraph: Discrete Reasoning over Visually-Rich Table-Text Documents via Semantic-Oriented Hierarchical Graphs

1 code implementation3 May 2023 Fengbin Zhu, Chao Wang, Fuli Feng, Zifeng Ren, Moxin Li, Tat-Seng Chua

Discrete reasoning over table-text documents (e. g., financial reports) gains increasing attention in recent two years.

Towards Complex Document Understanding By Discrete Reasoning

no code implementations25 Jul 2022 Fengbin Zhu, Wenqiang Lei, Fuli Feng, Chao Wang, Haozhou Zhang, Tat-Seng Chua

Document Visual Question Answering (VQA) aims to understand visually-rich documents to answer questions in natural language, which is an emerging research topic for both Natural Language Processing and Computer Vision.

document understanding Question Answering +1

RDU: A Region-based Approach to Form-style Document Understanding

no code implementations14 Jun 2022 Fengbin Zhu, Chao Wang, Wenqiang Lei, Ziyang Liu, Tat Seng Chua

Key Information Extraction (KIE) is aimed at extracting structured information (e. g. key-value pairs) from form-style documents (e. g. invoices), which makes an important step towards intelligent document understanding.

document understanding Form +6

TAT-QA: A Question Answering Benchmark on a Hybrid of Tabular and Textual Content in Finance

1 code implementation ACL 2021 Fengbin Zhu, Wenqiang Lei, Youcheng Huang, Chao Wang, Shuo Zhang, Jiancheng Lv, Fuli Feng, Tat-Seng Chua

In this work, we extract samples from real financial reports to build a new large-scale QA dataset containing both Tabular And Textual data, named TAT-QA, where numerical reasoning is usually required to infer the answer, such as addition, subtraction, multiplication, division, counting, comparison/sorting, and the compositions.

Question Answering

Retrieving and Reading: A Comprehensive Survey on Open-domain Question Answering

no code implementations4 Jan 2021 Fengbin Zhu, Wenqiang Lei, Chao Wang, Jianming Zheng, Soujanya Poria, Tat-Seng Chua

Open-domain Question Answering (OpenQA) is an important task in Natural Language Processing (NLP), which aims to answer a question in the form of natural language based on large-scale unstructured documents.

Machine Reading Comprehension Open-Domain Question Answering

Cannot find the paper you are looking for? You can Submit a new open access paper.