Search Results for author: Humam Alwassel

Found 9 papers, 6 papers with code

Exploring Missing Modality in Multimodal Egocentric Datasets

no code implementations21 Jan 2024 Merey Ramazanova, Alejandro Pardo, Humam Alwassel, Bernard Ghanem

Multimodal video understanding is crucial for analyzing egocentric videos, where integrating multiple sensory signals significantly enhances action recognition and moment localization.

Action Recognition Video Understanding

TSP: Temporally-Sensitive Pretraining of Video Encoders for Localization Tasks

1 code implementation23 Nov 2020 Humam Alwassel, Silvio Giancola, Bernard Ghanem

Extensive experiments show that using features trained with our novel pretraining strategy significantly improves the performance of recent state-of-the-art methods on three tasks: Temporal Action Localization, Action Proposal Generation, and Dense Video Captioning.

Action Classification Dense Video Captioning +2

Self-Supervised Learning by Cross-Modal Audio-Video Clustering

1 code implementation NeurIPS 2020 Humam Alwassel, Dhruv Mahajan, Bruno Korbar, Lorenzo Torresani, Bernard Ghanem, Du Tran

To the best of our knowledge, XDC is the first self-supervised learning method that outperforms large-scale fully-supervised pretraining for action recognition on the same architecture.

Audio Classification Clustering +5

MortonNet: Self-Supervised Learning of Local Features in 3D Point Clouds

1 code implementation30 Mar 2019 Ali Thabet, Humam Alwassel, Bernard Ghanem

In fact, we show how Morton features can be used to significantly improve performance (+3% for 2 popular semantic segmentation algorithms) in the task of semantic segmentation of point clouds on the challenging and large-scale S3DIS dataset.

Segmentation Self-Supervised Learning +1

The ActivityNet Large-Scale Activity Recognition Challenge 2018 Summary

no code implementations11 Aug 2018 Bernard Ghanem, Juan Carlos Niebles, Cees Snoek, Fabian Caba Heilbron, Humam Alwassel, Victor Escorcia, Ranjay Krishna, Shyamal Buch, Cuong Duc Dao

The guest tasks focused on complementary aspects of the activity recognition problem at large scale and involved three challenging and recently compiled datasets: the Kinetics-600 dataset from Google DeepMind, the AVA dataset from Berkeley and Google, and the Moments in Time dataset from MIT and IBM Research.

Activity Recognition

Diagnosing Error in Temporal Action Detectors

1 code implementation ECCV 2018 Humam Alwassel, Fabian Caba Heilbron, Victor Escorcia, Bernard Ghanem

Despite the recent progress in video understanding and the continuous rate of improvement in temporal action localization throughout the years, it is still unclear how far (or close?)

Temporal Action Localization Video Understanding

ActivityNet Challenge 2017 Summary

no code implementations22 Oct 2017 Bernard Ghanem, Juan Carlos Niebles, Cees Snoek, Fabian Caba Heilbron, Humam Alwassel, Ranjay Khrisna, Victor Escorcia, Kenji Hata, Shyamal Buch

The ActivityNet Large Scale Activity Recognition Challenge 2017 Summary: results and challenge participants papers.

Activity Recognition

Action Search: Spotting Actions in Videos and Its Application to Temporal Action Localization

1 code implementation ECCV 2018 Humam Alwassel, Fabian Caba Heilbron, Bernard Ghanem

To address this need, we propose the new problem of action spotting in video, which we define as finding a specific action in a video while observing a small portion of that video.

Action Spotting Temporal Action Localization

Cannot find the paper you are looking for? You can Submit a new open access paper.