1 code implementation • 14 Mar 2022 • Ruiwen Li, Zheda Mai, Chiheb Trabelsi, Zhibo Zhang, Jongseong Jang, Scott Sanner
In this paper, we propose TransCAM, a Conformer-based solution to WSSS that explicitly leverages the attention weights from the transformer branch of the Conformer to refine the CAM generated from the CNN branch.
Weakly supervised Semantic Segmentation Weakly-Supervised Semantic Segmentation
1 code implementation • 28 Nov 2021 • Zhibo Zhang, Jongseong Jang, Chiheb Trabelsi, Ruiwen Li, Scott Sanner, Yeonjeong Jeong, Dongsub Shim
Contrastive learning has led to substantial improvements in the quality of learned embedding representations for tasks such as image classification.
no code implementations • 29 May 2021 • Ruiwen Li, Zhibo Zhang, Jiani Li, Chiheb Trabelsi, Scott Sanner, Jongseong Jang, Yeonjeong Jeong, Dongsub Shim
Recent years have seen the introduction of a range of methods for post-hoc explainability of image classifier predictions.
1 code implementation • 25 Sep 2019 • Chiheb Trabelsi, Olexa Bilaniuk, Ousmane Dia, Ying Zhang, Mirco Ravanelli, Jonathan Binas, Negar Rostamzadeh, Christopher J Pal
Using the Wall Street Journal Dataset, we compare our phase-aware loss to several others that operate both in the time and frequency domains and demonstrate the effectiveness of our proposed signal extraction method and proposed loss.
no code implementations • NeurIPS Workshop Deep_Invers 2019 • Chiheb Trabelsi, Olexa Bilaniuk, Ousmane Dia, Ying Zhang, Mirco Ravanelli, Jonathan Binas, Negar Rostamzadeh, Christopher J Pal
Building on recent advances, we propose a new deep complex-valued method for signal retrieval and extraction in the frequency domain.
1 code implementation • 20 Jun 2018 • Titouan Parcollet, Ying Zhang, Mohamed Morchid, Chiheb Trabelsi, Georges Linarès, Renato De Mori, Yoshua Bengio
Quaternion numbers and quaternion neural networks have shown their efficiency to process multidimensional inputs as entities, to encode internal dependencies, and to solve many tasks with less learning parameters than real-valued models.
Ranked #19 on Speech Recognition on TIMIT
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
3 code implementations • ICLR 2019 • Titouan Parcollet, Mirco Ravanelli, Mohamed Morchid, Georges Linarès, Chiheb Trabelsi, Renato de Mori, Yoshua Bengio
Recurrent neural networks (RNNs) are powerful architectures to model sequential data, due to their capability to learn short and long-term dependencies between the basic elements of a sequence.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
no code implementations • ICML 2017 • Eugene Vorontsov, Chiheb Trabelsi, Samuel Kadoury, Chris Pal
We find that hard constraints on orthogonality can negatively affect the speed of convergence and model performance.
9 code implementations • ICLR 2018 • Chiheb Trabelsi, Olexa Bilaniuk, Ying Zhang, Dmitriy Serdyuk, Sandeep Subramanian, João Felipe Santos, Soroush Mehri, Negar Rostamzadeh, Yoshua Bengio, Christopher J. Pal
Despite their attractive properties and potential for opening up entirely new neural architectures, complex-valued deep neural networks have been marginalized due to the absence of the building blocks required to design such models.
Ranked #3 on Music Transcription on MusicNet
1 code implementation • 31 Jan 2017 • Eugene Vorontsov, Chiheb Trabelsi, Samuel Kadoury, Chris Pal
We find that hard constraints on orthogonality can negatively affect the speed of convergence and model performance.