no code implementations • 20 Mar 2024 • Zeyu Liu, Souvik Kundu, Anni Li, Junrui Wan, Lianghao Jiang, Peter Anthony Beerel
While compared in terms of runtime, AFLoRA can yield up to $1. 86\times$ improvement as opposed to similar PEFT alternatives.
1 code implementation • 20 Jan 2024 • Zeyu Liu, Gourav Datta, Anni Li, Peter Anthony Beerel
Moreover, we present a spiking version of this architecture, which introduces the benefit of states within the patch embedding and channel mixer modules while simultaneously reducing the computing complexity.
no code implementations • 29 Sep 2021 • Souvik Kundu, Peter Anthony Beerel, Sairam Sundaresan
In this paper, we present Fast Learnable Once-for-all Adversarial Training (FLOAT) which transforms the weight tensors without using extra layers, thereby incurring no significant increase in parameter count, training time, or network latency compared to a standard adversarial training.