no code implementations • ICLR 2019 • Shuhui Qu, Janghwan Lee, Wei Xiong, Wonhyouk Jang, Jie Wang
Since the generated samples simulate the low density area for each modal, the discriminator could directly detect anomalies from normal data.
no code implementations • 15 Sep 2023 • Qisen Cheng, Shuhui Qu, Janghwan Lee
Additionally, prediction with explanation serves as a regularizer, which improves the model's performance.
no code implementations • 11 Dec 2022 • Zheng Wang, Juncheng B Li, Shuhui Qu, Florian Metze, Emma Strubell
In this work, we incorporate exponentially decaying quantization-error-aware noise together with a learnable scale of task loss gradient to approximate the effect of a quantization operator.
no code implementations • 13 Oct 2022 • Zheng Wang, Juncheng B Li, Shuhui Qu, Florian Metze, Emma Strubell
Quantization is an effective technique to reduce memory footprint, inference latency, and power consumption of deep learning models.
no code implementations • 1 Jan 2021 • Juncheng B Li, Shuhui Qu, Xinjian Li, Emma Strubell, Florian Metze
Quantization of neural network parameters and activations has emerged as a successful approach to reducing the model size and inference time on hardware that sup-ports native low-precision arithmetic.
1 code implementation • 15 Nov 2020 • Juncheng B Li, Kaixin Ma, Shuhui Qu, Po-Yao Huang, Florian Metze
This work aims to study several key questions related to multimodal learning through the lens of adversarial noises: 1) The trade-off between early/middle/late fusion affecting its robustness and accuracy 2) How do different frequency/time domain features contribute to the robustness?
no code implementations • NeurIPS 2019 • Juncheng B. Li, Shuhui Qu, Xinjian Li, Joseph Szurley, J. Zico Kolter, Florian Metze
In this work, we target our attack on the wake-word detection system, jamming the model with some inconspicuous background music to deactivate the VAs while our audio adversary is present.
1 code implementation • 20 Mar 2017 • Juncheng Li, Wei Dai, Florian Metze, Shuhui Qu, Samarjit Das
On these features, we apply five models: Gaussian Mixture Model (GMM), Deep Neural Network (DNN), Recurrent Neural Network (RNN), Convolutional Deep Neural Net- work (CNN) and i-vector.
no code implementations • 29 Nov 2016 • Shuhui Qu, Juncheng Li, Wei Dai, Samarjit Das
Based on the procedure of log Mel-filter banks, we design a filter bank learning layer.
10 code implementations • 1 Oct 2016 • Wei Dai, Chia Dai, Shuhui Qu, Juncheng Li, Samarjit Das
Our CNNs, with up to 34 weight layers, are efficient to optimize over very long sequences (e. g., vector of size 32000), necessary for processing acoustic waveforms.