Search Results for author: Alejandro Pardo

Found 12 papers, 6 papers with code

MatchDiffusion: Training-free Generation of Match-cuts

1 code implementation27 Nov 2024 Alejandro Pardo, Fabio Pizzati, Tong Zhang, Alexander Pondaven, Philip Torr, Juan Camilo Perez, Bernard Ghanem

Match-cuts are powerful cinematic tools that create seamless transitions between scenes, delivering strong visual and metaphorical connections.

Denoising

Generative Timelines for Instructed Visual Assembly

no code implementations19 Nov 2024 Alejandro Pardo, Jui-Hsien Wang, Bernard Ghanem, Josef Sivic, Bryan Russell, Fabian Caba Heilbron

The objective of this work is to manipulate visual timelines (e. g. a video) through natural language instructions, making complex timeline editing tasks accessible to non-expert or potentially even disabled users.

Language Modelling

Compressed-Language Models for Understanding Compressed File Formats: a JPEG Exploration

no code implementations27 May 2024 Juan C. Pérez, Alejandro Pardo, Mattia Soldan, Hani Itani, Juan Leon-Alcazar, Bernard Ghanem

These results suggest that CLMs can understand the semantics of compressed data when directly operating on the byte streams of files produced by CFFs.

Combating Missing Modalities in Egocentric Videos at Test Time

no code implementations23 Apr 2024 Merey Ramazanova, Alejandro Pardo, Bernard Ghanem, Motasem Alfarra

Understanding videos that contain multiple modalities is crucial, especially in egocentric videos, where combining various sensory inputs significantly improves tasks like action recognition and moment localization.

Action Recognition Test-time Adaptation

Exploring Missing Modality in Multimodal Egocentric Datasets

no code implementations21 Jan 2024 Merey Ramazanova, Alejandro Pardo, Humam Alwassel, Bernard Ghanem

Multimodal video understanding is crucial for analyzing egocentric videos, where integrating multiple sensory signals significantly enhances action recognition and moment localization.

Action Recognition Video Understanding

Evaluation of Test-Time Adaptation Under Computational Time Constraints

1 code implementation10 Apr 2023 Motasem Alfarra, Hani Itani, Alejandro Pardo, Shyma Alhuwaider, Merey Ramazanova, Juan C. Pérez, Zhipeng Cai, Matthias Müller, Bernard Ghanem

To address this issue, we propose a more realistic evaluation protocol for TTA methods, where data is received in an online fashion from a constant-speed data stream, thereby accounting for the method's adaptation speed.

Test-time Adaptation

MovieCuts: A New Dataset and Benchmark for Cut Type Recognition

1 code implementation12 Sep 2021 Alejandro Pardo, Fabian Caba Heilbron, Juan León Alcázar, Ali Thabet, Bernard Ghanem

Advances in automatic Cut-type recognition can unleash new experiences in the video editing industry, such as movie analysis for education, video re-editing, virtual cinematography, machine-assisted trailer generation, machine-assisted video editing, among others.

Video Editing Vocal Bursts Type Prediction

Learning to Cut by Watching Movies

1 code implementation ICCV 2021 Alejandro Pardo, Fabian Caba Heilbron, Juan León Alcázar, Ali Thabet, Bernard Ghanem

Video content creation keeps growing at an incredible pace; yet, creating engaging stories remains challenging and requires non-trivial video editing expertise.

Contrastive Learning Video Editing

BAOD: Budget-Aware Object Detection

no code implementations10 Apr 2019 Alejandro Pardo, Mengmeng Xu, Ali Thabet, Pablo Arbelaez, Bernard Ghanem

We adopt a hybrid supervised learning framework to train the object detector from both these types of annotation.

Active Learning Object +2

Cannot find the paper you are looking for? You can Submit a new open access paper.