Search Results for author: Jiangxuan Long

Found 2 papers, 0 papers with code

Circuit Complexity Bounds for RoPE-based Transformer Architecture

no code implementations12 Nov 2024 Bo Chen, Xiaoyu Li, YIngyu Liang, Jiangxuan Long, Zhenmei Shi, Zhao Song

In this work, we establish a tighter circuit complexity bound for Transformers with $\mathsf{RoPE}$ attention.

Position

Beyond Linear Approximations: A Novel Pruning Approach for Attention Matrix

no code implementations15 Oct 2024 YIngyu Liang, Jiangxuan Long, Zhenmei Shi, Zhao Song, Yufa Zhou

Large Language Models (LLMs) have shown immense potential in enhancing various aspects of our daily lives, from conversational AI to search and AI assistants.

Cannot find the paper you are looking for? You can Submit a new open access paper.