no code implementations • 29 Oct 2024 • Chen Sun, Nolan Andrew Miller, Andrey Zhmoginov, Max Vladymyrov, Mark Sandler
What happens when a new piece of knowledge is introduced into the training data and how long does it last while a large language model (LM) continues to train?
no code implementations • 29 Aug 2024 • Rodrigo Diaz, Carlos De La Vega Martin, Mark Sandler
This paper presents an examination of State Space Models (SSM) and Koopman-based deep learning methods for modelling the dynamics of both linear and non-linear stiff strings.
no code implementations • 17 Aug 2024 • Gus Kristiansen, Mark Sandler, Andrey Zhmoginov, Nolan Miller, Anirudh Goyal, JIhwan Lee, Max Vladymyrov
In modern deep learning, the models are learned by applying gradient updates using an optimizer, which transforms the updates based on various statistics.
no code implementations • 21 Feb 2024 • Max Vladymyrov, Johannes von Oswald, Mark Sandler, Rong Ge
In this paper, we prove that each layer of a linear transformer maintains a weight vector for an implicit linear regression problem and can be interpreted as performing a variant of preconditioned gradient descent.
no code implementations • 7 Oct 2023 • Franco Caspe, Andrew McPherson, Mark Sandler
Tone Transfer is a novel deep-learning technique for interfacing a sound source with a synthesizer, transforming the timbre of audio excerpts while keeping their musical form content.
no code implementations • 11 Sep 2023 • Johannes von Oswald, Maximilian Schlegel, Alexander Meulemans, Seijin Kobayashi, Eyvind Niklasson, Nicolas Zucchet, Nino Scherrer, Nolan Miller, Mark Sandler, Blaise Agüera y Arcas, Max Vladymyrov, Razvan Pascanu, João Sacramento
Some autoregressive models exhibit in-context learning capabilities: being able to learn as an input sequence is processed, without undergoing any parameter changes, and without being explicitly trained to do so.
no code implementations • 11 Jan 2023 • Max Vladymyrov, Andrey Zhmoginov, Mark Sandler
We focus on the problem of learning without forgetting from multiple tasks arriving sequentially, where each task is defined using a few-shot episode of novel or already seen classes.
no code implementations • 5 Jan 2023 • Mark Sandler, Andrey Zhmoginov, Max Vladymyrov, Nolan Miller
In particular, for Exponential Moving Average (EMA) and Stochastic Weight Averaging we show that our proposed model matches the observed training trajectories on ImageNet.
no code implementations • CVPR 2023 • Andrey Zhmoginov, Mark Sandler, Nolan Miller, Gus Kristiansen, Max Vladymyrov
We study the effects of data and model architecture heterogeneity and the impact of the underlying communication graph topology on learning efficiency and show that our agents can significantly improve their performance compared to learning in isolation.
1 code implementation • 27 Oct 2022 • Rodrigo Diaz, Ben Hayes, Charalampos Saitis, György Fazekas, Mark Sandler
Physical models of rigid bodies are used for sound synthesis in applications from virtual environments to music production.
no code implementations • 12 Aug 2022 • Franco Caspe, Andrew McPherson, Mark Sandler
The training process involves a corpus of audio for supervision, and spectral reconstruction loss functions.
no code implementations • 15 Jul 2022 • Vinod Subramanian, Siddharth Gururani, Emmanouil Benetos, Mark Sandler
Loss-gradients are used to interpret the decision making process of deep learning models.
no code implementations • 10 Apr 2022 • Alejandro Delgado, Emir Demirel, Vinod Subramanian, Charalampos Saitis, Mark Sandler
Vocal Percussion Transcription (VPT) is concerned with the automatic detection and classification of vocal percussion sound events, allowing music creators and producers to sketch drum lines on the fly.
1 code implementation • 10 Apr 2022 • Alejandro Delgado, Charalampos Saitis, Emmanouil Benetos, Mark Sandler
Imitating musical instruments with the human voice is an efficient way of communicating ideas between music producers, from sketching melody lines to clarifying desired sonorities.
1 code implementation • CVPR 2022 • Mark Sandler, Andrey Zhmoginov, Max Vladymyrov, Andrew Jackson
In this paper we propose augmenting Vision Transformer models with learnable memory tokens.
1 code implementation • 11 Jan 2022 • Andrey Zhmoginov, Mark Sandler, Max Vladymyrov
In this work we propose a HyperTransformer, a Transformer-based model for supervised and semi-supervised few-shot learning that generates weights of a convolutional neural network (CNN) directly from support samples.
no code implementations • 29 Sep 2021 • Andrey Zhmoginov, Max Vladymyrov, Mark Sandler
In this work we propose a HyperTransformer, a transformer based model that generates all weights of a CNN model directly from the support samples.
no code implementations • 23 Jul 2021 • Andrey Zhmoginov, Dina Bashkirova, Mark Sandler
From practical perspective, our approach allows to: (a) reuse existing modules for learning new task by adjusting the computation order, (b) use it for unsupervised multi-source domain adaptation to illustrate that adaptation to unseen data can be achieved by only manipulating the order of pretrained modules, (c) show how our approach can be used to increase accuracy of existing architectures for image classification tasks such as ImageNet, without any parameter increase, by reusing the same block multiple times.
1 code implementation • 10 Apr 2021 • Mark Sandler, Max Vladymyrov, Andrey Zhmoginov, Nolan Miller, Andrew Jackson, Tom Madams, Blaise Aguera y Arcas
We show that classical gradient-based backpropagation in neural networks can be seen as a special case of a two-state network where one state is used for activations and another for gradients, with update rules derived from the chain rule.
no code implementations • 4 Jan 2021 • Keren Ye, Adriana Kovashka, Mark Sandler, Menglong Zhu, Andrew Howard, Marco Fornoni
In this paper we address the question: can task-specific detectors be trained and represented as a shared set of weights, plus a very small set of additional weights for each task?
no code implementations • 10 Dec 2020 • Liangchen Luo, Mark Sandler, Zi Lin, Andrey Zhmoginov, Andrew Howard
Knowledge distillation is one of the most popular and effective techniques for knowledge transfer, model compression and semi-supervised learning.
no code implementations • 11 Aug 2020 • Mark Sandler, Andrey Zhmoginov, Liangcheng Luo, Alexander Mordvintsev, Ettore Randazzo, Blaise Agúera y Arcas
The update rule is applied repeatedly in parallel to a large random subset of cells and after convergence is used to produce segmentation masks that are then back-propagated to learn the optimal update rules using standard gradient descent methods.
no code implementations • CVPR 2020 • Elad Eban, Yair Movshovitz-Attias, Hao Wu, Mark Sandler, Andrew Poon, Yerlan Idelbayev, Miguel A. Carreira-Perpinan
Despite the success of deep neural networks (DNNs), state-of-the-art models are too large to deploy on low-resource devices or common server configurations in which multiple models are held in memory.
no code implementations • 7 Sep 2019 • Mark Sandler, Jonathan Baccash, Andrey Zhmoginov, Andrew Howard
We explore the question of how the resolution of the input image ("input resolution") affects the performance of a neural network when compared to the resolution of the hidden layers ("internal resolution").
no code implementations • 22 Jul 2019 • Andrey Zhmoginov, Ian Fischer, Mark Sandler
We propose a new method for learning image attention masks in a semi-supervised setting based on the Information Bottleneck principle.
no code implementations • 4 Jul 2019 • Vinod Subramanian, Emmanouil Benetos, Ning Xu, SKoT McDonald, Mark Sandler
In addition, we show that the adversarial attacks are very effective across the different models.
1 code implementation • 8 May 2019 • Delia Fano Yela, Florian Thalmann, Vincenzo Nicosia, Dan Stowell, Mark Sandler
The empirical evidence suggests the proposed method for computation of visibility graphs offers an on-line computation solution at no additional computation time cost.
Data Structures and Algorithms
63 code implementations • ICCV 2019 • Andrew Howard, Mark Sandler, Grace Chu, Liang-Chieh Chen, Bo Chen, Mingxing Tan, Weijun Wang, Yukun Zhu, Ruoming Pang, Vijay Vasudevan, Quoc V. Le, Hartwig Adam
We achieve new state of the art results for mobile classification, detection and segmentation.
Ranked #9 on Dichotomous Image Segmentation on DIS-TE1
no code implementations • ICLR 2019 • Pramod Kaushik Mudrakarta, Mark Sandler, Andrey Zhmoginov, Andrew Howard
We introduce a novel method that enables parameter-efficient transfer and multitask learning.
1 code implementation • 5 Mar 2019 • Delia Fano Yela, Dan Stowell, Mark Sandler
We present experiments demonstrating the utility of this distance measure for real and synthesised audio data.
Sound Audio and Speech Processing
no code implementations • ICLR 2019 • Pramod Kaushik Mudrakarta, Mark Sandler, Andrey Zhmoginov, Andrew Howard
We introduce a novel method that enables parameter-efficient transfer and multi-task learning with deep neural networks.
28 code implementations • CVPR 2019 • Mingxing Tan, Bo Chen, Ruoming Pang, Vijay Vasudevan, Mark Sandler, Andrew Howard, Quoc V. Le
In this paper, we propose an automated mobile neural architecture search (MNAS) approach, which explicitly incorporate model latency into the main objective so that the search can identify a model that achieves a good trade-off between accuracy and latency.
Ranked #885 on Image Classification on ImageNet
4 code implementations • ECCV 2018 • Tien-Ju Yang, Andrew Howard, Bo Chen, Xiao Zhang, Alec Go, Mark Sandler, Vivienne Sze, Hartwig Adam
This work proposes an algorithm, called NetAdapt, that automatically adapts a pre-trained deep neural network to a mobile platform given a resource budget.
156 code implementations • CVPR 2018 • Mark Sandler, Andrew Howard, Menglong Zhu, Andrey Zhmoginov, Liang-Chieh Chen
In this paper we describe a new mobile architecture, MobileNetV2, that improves the state of the art performance of mobile models on multiple tasks and benchmarks as well as across a spectrum of different model sizes.
Ranked #7 on Retinal OCT Disease Classification on OCT2017
no code implementations • 8 Dec 2017 • Casey Chu, Andrey Zhmoginov, Mark Sandler
CycleGAN (Zhu et al. 2017) is one recent successful approach to learn a transformation between two image distributions.
2 code implementations • 13 Sep 2017 • Keunwoo Choi, György Fazekas, Kyunghyun Cho, Mark Sandler
Following their success in Computer Vision and other areas, deep learning techniques have recently become widely adopted in Music Information Retrieval (MIR) research.
1 code implementation • 6 Sep 2017 • Keunwoo Choi, György Fazekas, Kyunghyun Cho, Mark Sandler
In this paper, we empirically investigate the effect of audio preprocessing on music tagging with deep neural networks.
no code implementations • 7 Jun 2017 • Keunwoo Choi, George Fazekas, Kyunghyun Cho, Mark Sandler
The results highlight several important aspects of music tagging and neural networks.
3 code implementations • 27 Mar 2017 • Keunwoo Choi, György Fazekas, Mark Sandler, Kyunghyun Cho
In this paper, we present a transfer learning approach for music classification and regression tasks.
no code implementations • 21 Feb 2017 • Soravit Changpinyo, Mark Sandler, Andrey Zhmoginov
Deep convolutional networks are well-known for their high computational and memory demands.
13 code implementations • 14 Sep 2016 • Keunwoo Choi, George Fazekas, Mark Sandler, Kyunghyun Cho
We introduce a convolutional recurrent neural network (CRNN) for music tagging.
no code implementations • 17 Aug 2016 • Keunwoo Choi, George Fazekas, Brian McFee, Kyunghyun Cho, Mark Sandler
Descriptions are often provided along with recommendations to help users' discovery.
1 code implementation • 8 Jul 2016 • Keunwoo Choi, George Fazekas, Mark Sandler
Deep convolutional neural networks (CNNs) have been actively adopted in the field of music information retrieval, e. g. genre classification, mood detection, and chord recognition.
1 code implementation • 14 Jun 2016 • Andrey Zhmoginov, Mark Sandler
Deep neural networks have dramatically advanced the state of the art for many areas of machine learning.
no code implementations • 7 Jun 2016 • Keunwoo Choi, George Fazekas, Mark Sandler
We introduce a novel playlist generation algorithm that focuses on the quality of transitions using a recurrent neural network (RNN).
11 code implementations • 1 Jun 2016 • Keunwoo Choi, George Fazekas, Mark Sandler
We present a content-based automatic music tagging algorithm using fully convolutional neural networks (FCNs).
4 code implementations • 18 Apr 2016 • Keunwoo Choi, George Fazekas, Mark Sandler
In this paper, we introduce new methods and discuss results of text-based LSTM (Long Short-Term Memory) networks for automatic music composition.