no code implementations • 7 Sep 2023 • Taesik Gong, Josh Belanich, Krishna Somandepalli, Arsha Nagrani, Brian Eoff, Brendan Jou
Speech emotion recognition (SER) models typically rely on costly human-labeled data for training, making scaling methods to large speech datasets and nuanced emotion taxonomies difficult.
no code implementations • 24 Jun 2022 • Josh Belanich, Krishna Somandepalli, Brian Eoff, Brendan Jou
This technical report presents the modeling approaches used in our submission to the ICML Expressive Vocalizations Workshop & Competition multitask track (ExVo-MultiTask).
1 code implementation • ICLR 2022 • Asma Ghandeharioun, Been Kim, Chun-Liang Li, Brendan Jou, Brian Eoff, Rosalind W. Picard
Explaining deep learning model inferences is a promising venue for scientific understanding, improving safety, uncovering hidden biases, evaluating fairness, and beyond, as argued by many scholars.
no code implementations • 7 May 2021 • Mingda Zhang, Chun-Te Chu, Andrey Zhmoginov, Andrew Howard, Brendan Jou, Yukun Zhu, Li Zhang, Rebecca Hwa, Adriana Kovashka
With early termination, the average cost can be further reduced to 198M MAdds while maintaining accuracy of 80. 0% on ImageNet.
Ranked #661 on Image Classification on ImageNet
1 code implementation • 20 Sep 2019 • Asma Ghandeharioun, Brian Eoff, Brendan Jou, Rosalind W. Picard
Supporting model interpretability for complex phenomena where annotators can legitimately disagree, such as emotion recognition, is a challenging machine learning task.
3 code implementations • ICLR 2018 • Victor Campos, Brendan Jou, Xavier Giro-i-Nieto, Jordi Torres, Shih-Fu Chang
We introduce the Skip RNN model which extends existing RNN models by learning to skip state updates and shortens the effective size of the computational graph.
1 code implementation • 21 Aug 2017 • Delia Fernandez, Alejandro Woodward, Victor Campos, Xavier Giro-i-Nieto, Brendan Jou, Shih-Fu Chang
This work aims at disentangling the contributions of the `adjectives' and `nouns' in the visual prediction of ANPs.
no code implementations • 7 Jun 2016 • Nikolaos Pappas, Miriam Redi, Mercan Topkara, Brendan Jou, Hongyi Liu, Tao Chen, Shih-Fu Chang
The impact of culture in visual emotion perception has recently captured the attention of multimedia research.
no code implementations • 30 May 2016 • Brendan Jou, Shih-Fu Chang
In the original MVSO release, adjective-noun pair (ANP) detectors were trained for the six languages using an AlexNet-styled architecture by fine-tuning from DeepSentiBank.
2 code implementations • 12 Apr 2016 • Victor Campos, Brendan Jou, Xavier Giro-i-Nieto
Visual multimedia have become an inseparable part of our digital social lives, and they often capture moments tied with deep affections.
1 code implementation • 5 Apr 2016 • Brendan Jou, Shih-Fu Chang
We propose a novel extension of residual learning for deep networks that enables intuitive learning across multiple related tasks using cross-connections called cross-residuals.
1 code implementation • 20 Aug 2015 • Victor Campos, Amaia Salvador, Brendan Jou, Xavier Giró-i-Nieto
Visual media are powerful means of expressing emotions and sentiments.
no code implementations • 16 Aug 2015 • Brendan Jou, Tao Chen, Nikolaos Pappas, Miriam Redi, Mercan Topkara, Shih-Fu Chang
Our work expressly focuses on the uniqueness of culture and language in relation to human affect, specifically sentiment and emotion semantics, and how they manifest in social multimedia.
no code implementations • CVPR 2013 • Xin Guo, Dong Liu, Brendan Jou, Mojun Zhu, Anni Cai, Shih-Fu Chang
Object co-detection aims at simultaneous detection of objects of the same category from a pool of related images by exploiting consistent visual patterns present in candidate objects in the images.