Search Results for author: Yu-Shan Tai

Found 4 papers, 0 papers with code

MPTQ-ViT: Mixed-Precision Post-Training Quantization for Vision Transformer

no code implementations26 Jan 2024 Yu-Shan Tai, An-Yeu, Wu

However, without considering the asymmetry in activations and relying on hand-crafted settings, these methods often struggle to maintain performance under low-bit quantization.

Quantization

TSPTQ-ViT: Two-scaled post-training quantization for vision transformer

no code implementations22 May 2023 Yu-Shan Tai, Ming-Guang Lin, An-Yeu, Wu

Due to the non-normally distributed values after Softmax and GeLU, post-training quantization on ViTs results in severe accuracy degradation.

Quantization

Compression-aware Projection with Greedy Dimension Reduction for Convolutional Neural Network Activations

no code implementations17 Oct 2021 Yu-Shan Tai, Chieh-Fang Teng, Cheng-Yang Chang, An-Yeu Wu

Our test results show that the proposed methods effectively reduce 2. 91x~5. 97x memory access with negligible accuracy drop on MobileNetV2/ResNet18/VGG16.

Dimensionality Reduction

Cannot find the paper you are looking for? You can Submit a new open access paper.