Search Results for author: Alireza Ghaffari

Found 8 papers, 0 papers with code

Mitigating Outlier Activations in Low-Precision Fine-Tuning of Language Models

no code implementations14 Dec 2023 Alireza Ghaffari, Justin Yu, Mahsa Ghazvini Nejad, Masoud Asgharian, Boxing Chen, Vahid Partovi Nia

The benefit of using integers for outlier values is that it enables us to use operator tiling to avoid performing 16-bit integer matrix multiplication to address this problem effectively.

Statistical Hardware Design With Multi-model Active Learning

no code implementations14 Mar 2023 Alireza Ghaffari, Masoud Asgharian, Yvon Savaria

For instance, in our performance prediction setting, the proposed method needs 65% fewer samples to create the model, and in the design space exploration setting, our proposed method can find the best parameter settings by exploring less than 50 samples.

Active Learning Transfer Learning

On the Convergence of Stochastic Gradient Descent in Low-precision Number Formats

no code implementations4 Jan 2023 Matteo Cacciola, Antonio Frangioni, Masoud Asgharian, Alireza Ghaffari, Vahid Partovi Nia

Deep learning models are dominating almost all artificial intelligence tasks such as vision, text, and speech processing.

EuclidNets: An Alternative Operation for Efficient Inference of Deep Learning Models

no code implementations22 Dec 2022 Xinlin Li, Mariana Parazeres, Adam Oberman, Alireza Ghaffari, Masoud Asgharian, Vahid Partovi Nia

With the advent of deep learning application on edge devices, researchers actively try to optimize their deployments on low-power and restricted memory devices.

Quantization

Towards Fine-tuning Pre-trained Language Models with Integer Forward and Backward Propagation

no code implementations20 Sep 2022 Mohammadreza Tayaranian, Alireza Ghaffari, Marzieh S. Tahaei, Mehdi Rezagholizadeh, Masoud Asgharian, Vahid Partovi Nia

Previously researchers were focused on lower bit-width integer data types for the forward propagation of language models to save memory and computation.

Rethinking Pareto Frontier for Performance Evaluation of Deep Neural Networks

no code implementations18 Feb 2022 Vahid Partovi Nia, Alireza Ghaffari, Mahdi Zolnouri, Yvon Savaria

We propose to use a multi-dimensional Pareto frontier to re-define the efficiency measure of candidate deep learning models, where several variables such as training cost, inference latency, and accuracy play a relative role in defining a dominant model.

Benchmarking Image Classification

Cannot find the paper you are looking for? You can Submit a new open access paper.