Search Results for author: Cigdem Beyan

Found 14 papers, 8 papers with code

Leveraging Next-Active Objects for Context-Aware Anticipation in Egocentric Videos

no code implementations16 Aug 2023 Sanket Thakur, Cigdem Beyan, Pietro Morerio, Vittorio Murino, Alessio Del Bue

Compared to existing video modeling architectures for action anticipation, NAOGAT captures the relationship between objects and the global scene context in order to predict detections for the next active object and anticipate relevant future actions given these detections, leveraging the objects' dynamics to improve accuracy.

Action Anticipation Active Object Localization +3

Object-aware Gaze Target Detection

2 code implementations ICCV 2023 Francesco Tonini, Nicola Dall'Asen, Cigdem Beyan, Elisa Ricci

Gaze target detection aims to predict the image location where the person is looking and the probability that a gaze is out of the scene.

Object

Unsupervised Video Anomaly Detection with Diffusion Models Conditioned on Compact Motion Representations

1 code implementation4 Jul 2023 Anil Osman Tur, Nicola Dall'Asen, Cigdem Beyan, Elisa Ricci

This paper aims to address the unsupervised video anomaly detection (VAD) problem, which involves classifying each frame in a video as normal or abnormal, without any access to labels.

Anomaly Detection Video Anomaly Detection

Guided Attention for Next Active Object @ EGO4D STA Challenge

1 code implementation25 May 2023 Sanket Thakur, Cigdem Beyan, Pietro Morerio, Vittorio Murino, Alessio Del Bue

In this technical report, we describe the Guided-Attention mechanism based solution for the short-term anticipation (STA) challenge for the EGO4D challenge.

Object Short-term Object Interaction Anticipation

Enhancing Next Active Object-based Egocentric Action Anticipation with Guided Attention

1 code implementation22 May 2023 Sanket Thakur, Cigdem Beyan, Pietro Morerio, Vittorio Murino, Alessio Del Bue

To this end, we propose a novel approach that applies a guided attention mechanism between the objects, and the spatiotemporal features extracted from video clips, enhancing the motion and contextual information, and further decoding the object-centric and motion-centric information to address the problem of STA in egocentric videos.

Action Anticipation Object +1

Exploring Diffusion Models for Unsupervised Video Anomaly Detection

no code implementations12 Apr 2023 Anil Osman Tur, Nicola Dall'Asen, Cigdem Beyan, Elisa Ricci

This paper investigates the performance of diffusion models for video anomaly detection (VAD) within the most challenging but also the most operational scenario in which the data annotations are not used.

Anomaly Detection Video Anomaly Detection

Anticipating Next Active Objects for Egocentric Videos

no code implementations13 Feb 2023 Sanket Thakur, Cigdem Beyan, Pietro Morerio, Vittorio Murino, Alessio Del Bue

This paper addresses the problem of anticipating the next-active-object location in the future, for a given egocentric video clip where the contact might happen, before any action takes place.

Object

Multimodal Across Domains Gaze Target Detection

1 code implementation23 Aug 2022 Francesco Tonini, Cigdem Beyan, Elisa Ricci

This paper addresses the gaze target detection problem in single images captured from the third-person perspective.

Gaze Estimation Gaze Target Estimation

Co-Located Human-Human Interaction Analysis using Nonverbal Cues: A Survey

no code implementations20 Jul 2022 Cigdem Beyan, Alessandro Vinciarelli, Alessio Del Bue

Automated co-located human-human interaction analysis has been addressed by the use of nonverbal communication as measurable evidence of social and psychological phenomena.

Privacy Preserving

Estimating Presentation Competence using Multimodal Nonverbal Behavioral Cues

no code implementations6 May 2021 Ömer Sümer, Cigdem Beyan, Fabian Ruth, Olaf Kramer, Ulrich Trautwein, Enkelejda Kasneci

One approach that can promote efficient development of presentation competence is the automated analysis of human behavior during a speech based on visual and audio features and machine learning.

Cannot find the paper you are looking for? You can Submit a new open access paper.