Search Results for author: Joseph Hassoun

Found 6 papers, 4 papers with code

SaiT: Sparse Vision Transformers through Adaptive Token Pruning

1 code implementation11 Oct 2022 Ling Li, David Thorsley, Joseph Hassoun

Sparse adaptive image Transformer (SaiT) offers varying levels of model acceleration by merely changing the token sparsity on the fly.

Knowledge Distillation

MaiT: Leverage Attention Masks for More Efficient Image Transformers

no code implementations6 Jul 2022 Ling Li, Ali Shafiee Ardestani, Joseph Hassoun

Though image transformers have shown competitive results with convolutional neural networks in computer vision tasks, lacking inductive biases such as locality still poses problems in terms of model efficiency especially for embedded applications.

A Fast Post-Training Pruning Framework for Transformers

2 code implementations29 Mar 2022 Woosuk Kwon, Sehoon Kim, Michael W. Mahoney, Joseph Hassoun, Kurt Keutzer, Amir Gholami

To address this, we propose a fast post-training pruning framework for Transformers that does not require any retraining.

Learned Token Pruning for Transformers

1 code implementation2 Jul 2021 Sehoon Kim, Sheng Shen, David Thorsley, Amir Gholami, Woosuk Kwon, Joseph Hassoun, Kurt Keutzer

We extensively test the performance of LTP on GLUE tasks and show that our method outperforms the prior state-of-the-art token pruning methods by up to ~2. 5% higher accuracy with the same amount of FLOPs.

Sentence

Cannot find the paper you are looking for? You can Submit a new open access paper.