Search Results for author: Zuzana Jelčicová

Found 1 papers, 0 papers with code

Delta Keyword Transformer: Bringing Transformers to the Edge through Dynamically Pruned Multi-Head Self-Attention

no code implementations20 Mar 2022 Zuzana Jelčicová, Marian Verhelst

Moreover, a reduction of ~87-94% operations can be achieved when only degrading the accuracy by 1-4%, speeding up the multi-head self-attention inference by a factor of ~7. 5-16.

Keyword Spotting

Cannot find the paper you are looking for? You can Submit a new open access paper.