Search Results for author: Andrew Chang

Found 1 papers, 1 papers with code

HLSTransform: Energy-Efficient Llama 2 Inference on FPGAs Via High Level Synthesis

1 code implementation29 Apr 2024 Andy He, Darren Key, Mason Bulling, Andrew Chang, Skyler Shapiro, Everett Lee

Graphics Processing Units (GPUs) have become the leading hardware accelerator for deep learning applications and are used widely in training and inference of transformers; transformers have achieved state-of-the-art performance in many areas of machine learning and are especially used in most modern Large Language Models (LLMs).

Edge-computing

Cannot find the paper you are looking for? You can Submit a new open access paper.