no code implementations • 2 Feb 2024 • Mohammadreza Tayaranian, Seyyed Hasan Mozafari, James J. Clark, Brett Meyer, Warren Gross
In this work, we improve upon the inference latency of the state-of-the-art methods by removing the floating-point operations, which are associated with the GELU activation in Swin Transformer.
no code implementations • 24 Jan 2024 • Lulan Shen, Ali Edalati, Brett Meyer, Warren Gross, James J. Clark
This paper describes a simple yet effective technique for refining a pretrained classifier network.
no code implementations • 22 Jan 2024 • Lulan Shen, Ali Edalati, Brett Meyer, Warren Gross, James J. Clark
It is important to investigate the robustness of compressed networks in two types of data distribution shifts: domain shifts and adversarial perturbations.
no code implementations • 15 Sep 2022 • Ibtihel Amara, Maryam Ziaeefard, Brett H. Meyer, Warren Gross, James J. Clark
Knowledge distillation (KD) is an effective tool for compressing deep classification models for edge devices.
no code implementations • NeurIPS 2020 • Arash Ardakani, Amir Ardakani, Warren Gross
Therefore, our FSM-based model can learn extremely long-term dependencies as it requires 1/l memory storage during training compared to LSTMs, where l is the number of time steps.
no code implementations • 15 Sep 2020 • Nghia Doan, Seyyed Ali Hashemi, Warren Gross
In this paper we address the problem of selecting factor-graph permutations of polar codes under belief propagation (BP) decoding to significantly improve the error-correction performance of the code.
no code implementations • NeurIPS 2019 • Arash Ardakani, Zhengyun Ji, Amir Ardakani, Warren Gross
The emergence of XNOR networks seek to reduce the model size and computational cost of neural networks for their deployment on specialized hardware requiring real-time processes with limited hardware resources.