Search Results for author: Zhuoyuan Huo

Found 1 papers, 0 papers with code

TransformerFAM: Feedback attention is working memory

no code implementations14 Apr 2024 Dongseong Hwang, Weiran Wang, Zhuoyuan Huo, Khe Chai Sim, Pedro Moreno Mengibar

While Transformers have revolutionized deep learning, their quadratic attention complexity hinders their ability to process infinitely long inputs.

Cannot find the paper you are looking for? You can Submit a new open access paper.