Search Results for author: Aaron Chew

Found 1 papers, 0 papers with code

Efficiently Distilling LLMs for Edge Applications

no code implementations1 Apr 2024 Achintya Kundu, Fabian Lim, Aaron Chew, Laura Wynter, Penny Chong, Rhui Dih Lee

Supernet training of LLMs is of great interest in industrial applications as it confers the ability to produce a palette of smaller models at constant cost, regardless of the number of models (of different size / latency) produced.

Cannot find the paper you are looking for? You can Submit a new open access paper.