Search Results for author: Cameron R. Wolfe

Found 11 papers, 4 papers with code

Better Schedules for Low Precision Training of Deep Neural Networks

no code implementations4 Mar 2024 Cameron R. Wolfe, Anastasios Kyrillidis

From these experiments, we discover alternative CPT schedules that offer further improvements in training efficiency and model performance, as well as derive a set of best practices for choosing CPT schedules.

Node Classification Quantization +1

Cold Start Streaming Learning for Deep Networks

no code implementations9 Nov 2022 Cameron R. Wolfe, Anastasios Kyrillidis

To mitigate these shortcomings, we propose Cold Start Streaming Learning (CSSL), a simple, end-to-end approach for streaming learning with deep networks that uses a combination of replay and data augmentation to avoid catastrophic forgetting.

Data Augmentation

i-SpaSP: Structured Neural Pruning via Sparse Signal Recovery

1 code implementation7 Dec 2021 Cameron R. Wolfe, Anastasios Kyrillidis

We propose a novel, structured pruning algorithm for neural networks -- the iterative, Sparse Structured Pruning algorithm, dubbed as i-SpaSP.

How much pre-training is enough to discover a good subnetwork?

no code implementations31 Jul 2021 Cameron R. Wolfe, Fangshuo Liao, Qihan Wang, Junhyung Lyle Kim, Anastasios Kyrillidis

Aiming to mathematically analyze the amount of dense network pre-training needed for a pruned network to perform well, we discover a simple theoretical bound in the number of gradient descent pre-training iterations on a two-layer, fully-connected network, beyond which pruning via greedy forward selection [61] yields a subnetwork that achieves good training error.

Network Pruning

Exceeding the Limits of Visual-Linguistic Multi-Task Learning

no code implementations27 Jul 2021 Cameron R. Wolfe, Keld T. Lundgaard

By leveraging large amounts of product data collected across hundreds of live e-commerce websites, we construct 1000 unique classification tasks that share similarly-structured input data, comprised of both text and images.

Multi-Task Learning

ResIST: Layer-Wise Decomposition of ResNets for Distributed Training

no code implementations2 Jul 2021 Chen Dun, Cameron R. Wolfe, Christopher M. Jermaine, Anastasios Kyrillidis

Thus, ResIST reduces the per-iteration communication, memory, and time requirements of ResNet training to only a fraction of the requirements of full-model training.

GIST: Distributed Training for Large-Scale Graph Convolutional Networks

1 code implementation20 Feb 2021 Cameron R. Wolfe, Jingkang Yang, Arindam Chowdhury, Chen Dun, Artun Bayer, Santiago Segarra, Anastasios Kyrillidis

The graph convolutional network (GCN) is a go-to solution for machine learning on graphs, but its training is notoriously difficult to scale both in terms of graph size and the number of model parameters.

BIG-bench Machine Learning Graph Sampling

E-Stitchup: Data Augmentation for Pre-Trained Embeddings

no code implementations28 Nov 2019 Cameron R. Wolfe, Keld T. Lundgaard

In this work, we propose data augmentation methods for embeddings from pre-trained deep learning models that take a weighted combination of a pair of input embeddings, as inspired by Mixup, and combine such augmentation with extra label softening.

Data Augmentation General Classification +1

Distributed Learning of Deep Neural Networks using Independent Subnet Training

2 code implementations4 Oct 2019 Binhang Yuan, Cameron R. Wolfe, Chen Dun, Yuxin Tang, Anastasios Kyrillidis, Christopher M. Jermaine

These properties of IST can cope with issues due to distributed data, slow interconnects, or limited device memory, making IST a suitable approach for cases of mandatory distribution.

BIG-bench Machine Learning Image Classification +2

Functional Generative Design of Mechanisms with Recurrent Neural Networks and Novelty Search

no code implementations25 Mar 2019 Cameron R. Wolfe, Cem C. Tutum, Risto Miikkulainen

However, while static designs are easily produced with 3D printing, functional designs with moving parts are more difficult to generate: The search space is too high-dimensional, the resolution of the 3D-printed parts is not adequate, and it is difficult to predict the physical behavior of imperfect 3D-printed mechanisms.

Cannot find the paper you are looking for? You can Submit a new open access paper.