Search Results for author: Mark Sandler

Found 44 papers, 19 papers with code

Linear Transformers are Versatile In-Context Learners

no code implementations21 Feb 2024 Max Vladymyrov, Johannes von Oswald, Mark Sandler, Rong Ge

Recent research has demonstrated that transformers, particularly linear attention models, implicitly execute gradient-descent-like algorithms on data provided in-context during their forward inference step.

FM Tone Transfer with Envelope Learning

no code implementations7 Oct 2023 Franco Caspe, Andrew McPherson, Mark Sandler

Tone Transfer is a novel deep-learning technique for interfacing a sound source with a synthesizer, transforming the timbre of audio excerpts while keeping their musical form content.

Continual Few-Shot Learning Using HyperTransformers

no code implementations11 Jan 2023 Max Vladymyrov, Andrey Zhmoginov, Mark Sandler

We focus on the problem of learning without forgetting from multiple tasks arriving sequentially, where each task is defined using a few-shot episode of novel or already seen classes.

Class Incremental Learning continual few-shot learning +2

Training trajectories, mini-batch losses and the curious role of the learning rate

no code implementations5 Jan 2023 Mark Sandler, Andrey Zhmoginov, Max Vladymyrov, Nolan Miller

In particular, for Exponential Moving Average (EMA) and Stochastic Weight Averaging we show that our proposed model matches the observed training trajectories on ImageNet.

Decentralized Learning with Multi-Headed Distillation

no code implementations CVPR 2023 Andrey Zhmoginov, Mark Sandler, Nolan Miller, Gus Kristiansen, Max Vladymyrov

We study the effects of data and model architecture heterogeneity and the impact of the underlying communication graph topology on learning efficiency and show that our agents can significantly improve their performance compared to learning in isolation.

Rigid-Body Sound Synthesis with Differentiable Modal Resonators

1 code implementation27 Oct 2022 Rodrigo Diaz, Ben Hayes, Charalampos Saitis, György Fazekas, Mark Sandler

Physical models of rigid bodies are used for sound synthesis in applications from virtual environments to music production.

DDX7: Differentiable FM Synthesis of Musical Instrument Sounds

no code implementations12 Aug 2022 Franco Caspe, Andrew McPherson, Mark Sandler

The training process involves a corpus of audio for supervision, and spectral reconstruction loss functions.

Continuous Control Resynthesis +1

Deep Embeddings for Robust User-Based Amateur Vocal Percussion Classification

no code implementations10 Apr 2022 Alejandro Delgado, Emir Demirel, Vinod Subramanian, Charalampos Saitis, Mark Sandler

Vocal Percussion Transcription (VPT) is concerned with the automatic detection and classification of vocal percussion sound events, allowing music creators and producers to sketch drum lines on the fly.

Classification feature selection +2

Deep Conditional Representation Learning for Drum Sample Retrieval by Vocalisation

1 code implementation10 Apr 2022 Alejandro Delgado, Charalampos Saitis, Emmanouil Benetos, Mark Sandler

Imitating musical instruments with the human voice is an efficient way of communicating ideas between music producers, from sketching melody lines to clarifying desired sonorities.

Representation Learning Retrieval

Fine-tuning Image Transformers using Learnable Memory

1 code implementation CVPR 2022 Mark Sandler, Andrey Zhmoginov, Max Vladymyrov, Andrew Jackson

In this paper we propose augmenting Vision Transformer models with learnable memory tokens.

HyperTransformer: Model Generation for Supervised and Semi-Supervised Few-Shot Learning

1 code implementation11 Jan 2022 Andrey Zhmoginov, Mark Sandler, Max Vladymyrov

In this work we propose a HyperTransformer, a Transformer-based model for supervised and semi-supervised few-shot learning that generates weights of a convolutional neural network (CNN) directly from support samples.

Few-Shot Image Classification Few-Shot Learning

HyperTransformer: Attention-Based CNN Model Generation from Few Samples

no code implementations29 Sep 2021 Andrey Zhmoginov, Max Vladymyrov, Mark Sandler

In this work we propose a HyperTransformer, a transformer based model that generates all weights of a CNN model directly from the support samples.

Few-Shot Learning

Compositional Models: Multi-Task Learning and Knowledge Transfer with Modular Networks

no code implementations23 Jul 2021 Andrey Zhmoginov, Dina Bashkirova, Mark Sandler

From practical perspective, our approach allows to: (a) reuse existing modules for learning new task by adjusting the computation order, (b) use it for unsupervised multi-source domain adaptation to illustrate that adaptation to unseen data can be achieved by only manipulating the order of pretrained modules, (c) show how our approach can be used to increase accuracy of existing architectures for image classification tasks such as ImageNet, without any parameter increase, by reusing the same block multiple times.

Domain Adaptation Image Classification +1

Meta-Learning Bidirectional Update Rules

1 code implementation10 Apr 2021 Mark Sandler, Max Vladymyrov, Andrey Zhmoginov, Nolan Miller, Andrew Jackson, Tom Madams, Blaise Aguera y Arcas

We show that classical gradient-based backpropagation in neural networks can be seen as a special case of a two-state network where one state is used for activations and another for gradients, with update rules derived from the chain rule.

Meta-Learning

SpotPatch: Parameter-Efficient Transfer Learning for Mobile Object Detection

no code implementations4 Jan 2021 Keren Ye, Adriana Kovashka, Mark Sandler, Menglong Zhu, Andrew Howard, Marco Fornoni

In this paper we address the question: can task-specific detectors be trained and represented as a shared set of weights, plus a very small set of additional weights for each task?

Object object-detection +2

Large-Scale Generative Data-Free Distillation

no code implementations10 Dec 2020 Liangchen Luo, Mark Sandler, Zi Lin, Andrey Zhmoginov, Andrew Howard

Knowledge distillation is one of the most popular and effective techniques for knowledge transfer, model compression and semi-supervised learning.

Knowledge Distillation Model Compression +1

Image segmentation via Cellular Automata

no code implementations11 Aug 2020 Mark Sandler, Andrey Zhmoginov, Liangcheng Luo, Alexander Mordvintsev, Ettore Randazzo, Blaise Agúera y Arcas

The update rule is applied repeatedly in parallel to a large random subset of cells and after convergence is used to produce segmentation masks that are then back-propagated to learn the optimal update rules using standard gradient descent methods.

Image Segmentation Segmentation +1

Structured Multi-Hashing for Model Compression

no code implementations CVPR 2020 Elad Eban, Yair Movshovitz-Attias, Hao Wu, Mark Sandler, Andrew Poon, Yerlan Idelbayev, Miguel A. Carreira-Perpinan

Despite the success of deep neural networks (DNNs), state-of-the-art models are too large to deploy on low-resource devices or common server configurations in which multiple models are held in memory.

Model Compression

Non-discriminative data or weak model? On the relative importance of data and model resolution

no code implementations7 Sep 2019 Mark Sandler, Jonathan Baccash, Andrey Zhmoginov, Andrew Howard

We explore the question of how the resolution of the input image ("input resolution") affects the performance of a neural network when compared to the resolution of the hidden layers ("internal resolution").

Information-Bottleneck Approach to Salient Region Discovery

no code implementations22 Jul 2019 Andrey Zhmoginov, Ian Fischer, Mark Sandler

We propose a new method for learning image attention masks in a semi-supervised setting based on the Information Bottleneck principle.

Efficient On-line Computation of Visibility Graphs

1 code implementation8 May 2019 Delia Fano Yela, Florian Thalmann, Vincenzo Nicosia, Dan Stowell, Mark Sandler

The empirical evidence suggests the proposed method for computation of visibility graphs offers an on-line computation solution at no additional computation time cost.

Data Structures and Algorithms

Visibility graphs for robust harmonic similarity measures between audio spectra

1 code implementation5 Mar 2019 Delia Fano Yela, Dan Stowell, Mark Sandler

We present experiments demonstrating the utility of this distance measure for real and synthesised audio data.

Sound Audio and Speech Processing

MnasNet: Platform-Aware Neural Architecture Search for Mobile

28 code implementations CVPR 2019 Mingxing Tan, Bo Chen, Ruoming Pang, Vijay Vasudevan, Mark Sandler, Andrew Howard, Quoc V. Le

In this paper, we propose an automated mobile neural architecture search (MNAS) approach, which explicitly incorporate model latency into the main objective so that the search can identify a model that achieves a good trade-off between accuracy and latency.

Ranked #832 on Image Classification on ImageNet (using extra training data)

Image Classification Neural Architecture Search +2

NetAdapt: Platform-Aware Neural Network Adaptation for Mobile Applications

4 code implementations ECCV 2018 Tien-Ju Yang, Andrew Howard, Bo Chen, Xiao Zhang, Alec Go, Mark Sandler, Vivienne Sze, Hartwig Adam

This work proposes an algorithm, called NetAdapt, that automatically adapts a pre-trained deep neural network to a mobile platform given a resource budget.

Image Classification

MobileNetV2: Inverted Residuals and Linear Bottlenecks

148 code implementations CVPR 2018 Mark Sandler, Andrew Howard, Menglong Zhu, Andrey Zhmoginov, Liang-Chieh Chen

In this paper we describe a new mobile architecture, MobileNetV2, that improves the state of the art performance of mobile models on multiple tasks and benchmarks as well as across a spectrum of different model sizes.

Image Classification Image Segmentation +4

CycleGAN, a Master of Steganography

no code implementations8 Dec 2017 Casey Chu, Andrey Zhmoginov, Mark Sandler

CycleGAN (Zhu et al. 2017) is one recent successful approach to learn a transformation between two image distributions.

A Tutorial on Deep Learning for Music Information Retrieval

2 code implementations13 Sep 2017 Keunwoo Choi, György Fazekas, Kyunghyun Cho, Mark Sandler

Following their success in Computer Vision and other areas, deep learning techniques have recently become widely adopted in Music Information Retrieval (MIR) research.

Information Retrieval Music Information Retrieval +2

A Comparison of Audio Signal Preprocessing Methods for Deep Neural Networks on Music Tagging

1 code implementation6 Sep 2017 Keunwoo Choi, György Fazekas, Kyunghyun Cho, Mark Sandler

In this paper, we empirically investigate the effect of audio preprocessing on music tagging with deep neural networks.

Music Tagging

Transfer learning for music classification and regression tasks

3 code implementations27 Mar 2017 Keunwoo Choi, György Fazekas, Mark Sandler, Kyunghyun Cho

In this paper, we present a transfer learning approach for music classification and regression tasks.

Classification General Classification +4

The Power of Sparsity in Convolutional Neural Networks

no code implementations21 Feb 2017 Soravit Changpinyo, Mark Sandler, Andrey Zhmoginov

Deep convolutional networks are well-known for their high computational and memory demands.

Explaining Deep Convolutional Neural Networks on Music Classification

1 code implementation8 Jul 2016 Keunwoo Choi, George Fazekas, Mark Sandler

Deep convolutional neural networks (CNNs) have been actively adopted in the field of music information retrieval, e. g. genre classification, mood detection, and chord recognition.

Chord Recognition Classification +6

Inverting face embeddings with convolutional neural networks

1 code implementation14 Jun 2016 Andrey Zhmoginov, Mark Sandler

Deep neural networks have dramatically advanced the state of the art for many areas of machine learning.

Face Transfer

Towards Playlist Generation Algorithms Using RNNs Trained on Within-Track Transitions

no code implementations7 Jun 2016 Keunwoo Choi, George Fazekas, Mark Sandler

We introduce a novel playlist generation algorithm that focuses on the quality of transitions using a recurrent neural network (RNN).

Automatic tagging using deep convolutional neural networks

11 code implementations1 Jun 2016 Keunwoo Choi, George Fazekas, Mark Sandler

We present a content-based automatic music tagging algorithm using fully convolutional neural networks (FCNs).

Music Tagging

Text-based LSTM networks for Automatic Music Composition

4 code implementations18 Apr 2016 Keunwoo Choi, George Fazekas, Mark Sandler

In this paper, we introduce new methods and discuss results of text-based LSTM (Long Short-Term Memory) networks for automatic music composition.

Cannot find the paper you are looking for? You can Submit a new open access paper.