Search Results for author: Suyuchen Wang

Found 12 papers, 6 papers with code

FinSage: A Multi-aspect RAG System for Financial Filings Question Answering

no code implementations20 Apr 2025 Xinyu Wang, Jijun Chi, Zhenghan Tai, Tung Sum Thomas Kwok, Muzhi Li, Zhuhong Li, Hailin He, Yuchen Hua, Peng Lu, Suyuchen Wang, Yihong Wu, Jerry Huang, Ling Zhou

Leveraging large language models in real-world settings often entails a need to utilize domain-specific data and tools in order to follow the complex regulations that need to be followed for acceptable use.

Question Answering RAG +2

GraphOmni: A Comprehensive and Extendable Benchmark Framework for Large Language Models on Graph-theoretic Tasks

no code implementations17 Apr 2025 Hao Xu, Xiangru Jian, Xinjian Zhao, Wei Pang, Chao Zhang, Suyuchen Wang, Qixin Zhang, Joao Monteiro, Qiuzhuang Sun, Tianshu Yu

In this paper, we presented GraphOmni, a comprehensive benchmark framework for systematically evaluating the graph reasoning capabilities of LLMs.

R$^3$Mem: Bridging Memory Retention and Retrieval via Reversible Compression

no code implementations21 Feb 2025 Xiaoqiang Wang, Suyuchen Wang, Yun Zhu, Bang Liu

For retrieval, R$^3$Mem employs a reversible architecture, reconstructing raw data by invoking the model backward with compressed information.

Language Modeling Language Modelling +2

FACT: Examining the Effectiveness of Iterative Context Rewriting for Multi-fact Retrieval

no code implementations28 Oct 2024 Jinlin Wang, Suyuchen Wang, Ziwen Xia, Sirui Hong, Yun Zhu, Bang Liu, Chenglin Wu

Large Language Models (LLMs) are proficient at retrieving single facts from extended contexts, yet they struggle with tasks requiring the simultaneous retrieval of multiple facts, especially during generation.

Retrieval

LongRecipe: Recipe for Efficient Long Context Generalization in Large Language Models

1 code implementation31 Aug 2024 Zhiyuan Hu, Yuliang Liu, Jinman Zhao, Suyuchen Wang, Yan Wang, Wei Shen, Qing Gu, Anh Tuan Luu, See-Kiong Ng, Zhiwei Jiang, Bryan Hooi

Large language models (LLMs) face significant challenges in handling long-context tasks because of their limited effective context window size during pretraining, which restricts their ability to generalize over extended sequences.

8k

MAP: Low-compute Model Merging with Amortized Pareto Fronts via Quadratic Approximation

1 code implementation11 Jun 2024 Lu Li, Tianyu Zhang, Zhiqi Bu, Suyuchen Wang, Huan He, Jie Fu, Yonghui Wu, Jiang Bian, Yong Chen, Yoshua Bengio

MAP efficiently identifies a Pareto set of scaling coefficients for merging multiple models, reflecting the trade-offs involved.

VCR: A Task for Pixel-Level Complex Reasoning in Vision Language Models via Restoring Occluded Text

1 code implementation10 Jun 2024 Tianyu Zhang, Suyuchen Wang, Lu Li, Ge Zhang, Perouz Taslakian, Sai Rajeswar, Jie Fu, Bang Liu, Yoshua Bengio

This task stems from the observation that text embedded in images is intrinsically different from common visual elements and natural language due to the need to align the modalities of vision, text, and text embedded in images.

Language Modeling Language Modelling +4

Resonance RoPE: Improving Context Length Generalization of Large Language Models

1 code implementation29 Feb 2024 Suyuchen Wang, Ivan Kobyzev, Peng Lu, Mehdi Rezagholizadeh, Bang Liu

This paper addresses the challenge of train-short-test-long (TSTL) scenarios in Large Language Models (LLMs) equipped with Rotary Position Embedding (RoPE), where models pre-trained on shorter sequences face difficulty with out-of-distribution (OOD) token positions in longer sequences.

Language Modeling Language Modelling +1

Cannot find the paper you are looking for? You can Submit a new open access paper.