Search Results for author: Brett H. Meyer

Found 9 papers, 2 papers with code

SSS3D: Fast Neural Architecture Search For Efficient Three-Dimensional Semantic Segmentation

no code implementations21 Apr 2023 Olivier Therrien, Marihan Amein, Zhuoran Xiong, Warren J. Gross, Brett H. Meyer

We present SSS3D, a fast multi-objective NAS framework designed to find computationally efficient 3D semantic scene segmentation networks.

Neural Architecture Search Scene Segmentation

BD-KD: Balancing the Divergences for Online Knowledge Distillation

no code implementations25 Dec 2022 Ibtihel Amara, Nazanin Sepahvand, Brett H. Meyer, Warren J. Gross, James J. Clark

We show that adaptively balancing between the reverse and forward divergences shifts the focus of the training strategy to the compact student network without limiting the teacher network's learning process.

Knowledge Distillation Model Compression +1

CES-KD: Curriculum-based Expert Selection for Guided Knowledge Distillation

no code implementations15 Sep 2022 Ibtihel Amara, Maryam Ziaeefard, Brett H. Meyer, Warren Gross, James J. Clark

Knowledge distillation (KD) is an effective tool for compressing deep classification models for edge devices.

Knowledge Distillation

Efficient Fine-Tuning of Compressed Language Models with Learners

no code implementations3 Aug 2022 Danilo Vucetic, Mohammadreza Tayaranian, Maryam Ziaeefard, James J. Clark, Brett H. Meyer, Warren J. Gross

We introduce Learner modules and priming, novel methods for fine-tuning that exploit the overparameterization of pre-trained language models to gain benefits in convergence speed and resource utilization.

CoLA Navigate

Efficient Fine-Tuning of BERT Models on the Edge

no code implementations3 May 2022 Danilo Vucetic, Mohammadreza Tayaranian, Maryam Ziaeefard, James J. Clark, Brett H. Meyer, Warren J. Gross

FAR reduces fine-tuning time on the DistilBERT model and CoLA dataset by 30%, and time spent on memory operations by 47%.

CoLA

Surprisal-Triggered Conditional Computation with Neural Networks

1 code implementation2 Jun 2020 Loren Lugosch, Derek Nowrouzezahrai, Brett H. Meyer

The surprisal of the input, measured as the negative log-likelihood of the current observation according to the autoregressive model, is used as a measure of input difficulty.

speech-recognition Speech Recognition

Learning Recurrent Binary/Ternary Weights

1 code implementation ICLR 2019 Arash Ardakani, Zhengyun Ji, Sean C. Smithson, Brett H. Meyer, Warren J. Gross

On the software side, we evaluate the performance (in terms of accuracy) of our method using long short-term memories (LSTMs) on various sequential models including sequence classification and language modeling.

Language Modelling

Neural Networks Designing Neural Networks: Multi-Objective Hyper-Parameter Optimization

no code implementations7 Nov 2016 Sean C. Smithson, Guang Yang, Warren J. Gross, Brett H. Meyer

The method is evaluated on the MNIST and CIFAR-10 image datasets, optimizing for both recognition accuracy and computational complexity.

BIG-bench Machine Learning Image Classification +2

Cannot find the paper you are looking for? You can Submit a new open access paper.