no code implementations • 20 Sep 2024 • Aditya Kommineni, Digbalay Bose, Tiantian Feng, So Hyun Kim, Helen Tager-Flusberg, Somer Bishop, Catherine Lord, Sudarsana Kadiri, Shrikanth Narayanan
Objective analyses of these videos could provide clinicians and researchers with nuanced insights into the behavior of children with Autism Spectrum Disorder.
no code implementations • 14 Feb 2024 • Tiantian Feng, Daniel Yang, Digbalay Bose, Shrikanth Narayanan
Specifically, we propose a simple but effective multi-modal learning framework GTI-MM to enhance the data efficiency and model robustness against missing visual modality by imputing the missing data with generative transformers.
no code implementations • 18 Sep 2023 • Yoonsoo Nam, Adam Lehavi, Daniel Yang, Digbalay Bose, Swabha Swayamdipta, Shrikanth Narayanan
Video summarization remains a huge challenge in computer vision due to the size of the input videos to be summarized.
no code implementations • 27 Aug 2023 • Digbalay Bose, Rajat Hebbar, Tiantian Feng, Krishna Somandepalli, Anfeng Xu, Shrikanth Narayanan
Advertisement videos (ads) play an integral part in the domain of Internet e-commerce as they amplify the reach of particular products to a broad audience or can serve as a medium to raise awareness about specific issues through concise narrative structures.
no code implementations • 15 Jun 2023 • Tiantian Feng, Digbalay Bose, Tuo Zhang, Rajat Hebbar, Anil Ramakrishna, Rahul Gupta, Mi Zhang, Salman Avestimehr, Shrikanth Narayanan
In order to facilitate the research in multimodal FL, we introduce FedMultimodal, the first FL benchmark for multimodal learning covering five representative multimodal applications from ten commonly used datasets with a total of eight unique modalities.
no code implementations • 17 Apr 2023 • Kleanthis Avramidis, Kranti Adsul, Digbalay Bose, Shrikanth Narayanan
This paper presents the approach and results of USC SAIL's submission to the Signal Processing Grand Challenge 2023 - e-Prevention (Task 2), on detecting relapses in psychotic patients.
1 code implementation • 13 Mar 2023 • Digbalay Bose, Rajat Hebbar, Krishna Somandepalli, Shrikanth Narayanan
The process of human affect understanding involves the ability to infer person specific emotional states from various sources including images, speech, and language.
1 code implementation • 14 Feb 2023 • Rajat Hebbar, Digbalay Bose, Krishna Somandepalli, Veena Vijai, Shrikanth Narayanan
In this work, we present a dataset of audio events called Subtitle-Aligned Movie Sounds (SAM-S).
1 code implementation • 28 Oct 2022 • Kleanthis Avramidis, Tiantian Feng, Digbalay Bose, Shrikanth Narayanan
Detecting unsafe driving states, such as stress, drowsiness, and fatigue, is an important component of ensuring driving safety and an essential prerequisite for automatic intervention systems in vehicles.
1 code implementation • 20 Oct 2022 • Digbalay Bose, Rajat Hebbar, Krishna Somandepalli, Haoyang Zhang, Yin Cui, Kree Cole-McLaughlin, Huisheng Wang, Shrikanth Narayanan
Longform media such as movies have complex narrative structures, with events spanning a rich variety of ambient visual scenes.
no code implementations • 13 Oct 2021 • Digbalay Bose, Krishna Somandepalli, Souvik Kundu, Rimita Lahiri, Jonathan Gratch, Shrikanth Narayanan
Computational modeling of the emotions evoked by art in humans is a challenging problem because of the subjective and nuanced nature of art and affective signals.
no code implementations • 11 Oct 2021 • Justin Olah, Sabyasachee Baruah, Digbalay Bose, Shrikanth Narayanan
Emotion recognition from text is a challenging task due to diverse emotion taxonomies, lack of reliable labeled data in different domains, and highly subjective annotation standards.
no code implementations • 14 Dec 2013 • Srinjoy Ganguly, Digbalay Bose, Amit Konar
We also examine the efficacy of the proposed scheme by analyzing its performance on image segmentation examples and comparing it with the classical Fuzzy C-means clustering algorithm.