no code implementations • 12 Jun 2024 • Michele Mazzamuto, Antonino Furnari, Giovanni Maria Farinella
In this paper, we address the challenge of unsupervised mistake detection in egocentric procedural video through the analysis of gaze signals.
1 code implementation • 3 Jun 2024 • Lorenzo Mur-Labadia, Ruben Martinez-Cantin, Josechu Guerrero, Giovanni Maria Farinella, Antonino Furnari
Short-Term object-interaction Anticipation consists of detecting the location of the next-active objects, the noun and verb categories of the interaction, and the time to contact from the observation of egocentric video.
1 code implementation • 3 Jun 2024 • Luigi Seminara, Giovanni Maria Farinella, Antonino Furnari
Task graphs learned with our approach are also shown to significantly enhance online mistake detection in procedural egocentric videos, achieving notable gains of +19. 8% and +7. 5% on the Assembly101 and EPIC-Tent datasets.
1 code implementation • CVPR 2024 • Alessandro Flaborea, Guido Maria D'Amely di Melendugno, Leonardo Plini, Luca Scofano, Edoardo De Matteis, Antonino Furnari, Giovanni Maria Farinella, Fabio Galasso
We propose PREGO, the first online one-class classification model for mistake detection in PRocedural EGOcentric videos.
1 code implementation • CVPR 2024 • Ivan Rodin, Antonino Furnari, Kyle Min, Subarna Tripathi, Giovanni Maria Farinella
We present Egocentric Action Scene Graphs (EASGs), a new representation for long-form understanding of egocentric videos.
no code implementations • 5 Dec 2023 • Camillo Quattrocchi, Antonino Furnari, Daniele Di Mauro, Mario Valerio Giuffrida, Giovanni Maria Farinella
Instead, we propose a novel methodology which performs the adaptation leveraging existing labeled exocentric videos and a new set of unlabeled, synchronized exocentric-egocentric video pairs, for which temporal action segmentation annotations do not need to be collected.
1 code implementation • 5 Dec 2023 • Rosario Leonardi, Antonino Furnari, Francesco Ragusa, Giovanni Maria Farinella
In this study, we investigate the effectiveness of synthetic data in enhancing egocentric hand-object interaction detection.
2 code implementations • CVPR 2024 • Kristen Grauman, Andrew Westbury, Lorenzo Torresani, Kris Kitani, Jitendra Malik, Triantafyllos Afouras, Kumar Ashutosh, Vijay Baiyya, Siddhant Bansal, Bikram Boote, Eugene Byrne, Zach Chavis, Joya Chen, Feng Cheng, Fu-Jen Chu, Sean Crane, Avijit Dasgupta, Jing Dong, Maria Escobar, Cristhian Forigua, Abrham Gebreselasie, Sanjay Haresh, Jing Huang, Md Mohaiminul Islam, Suyog Jain, Rawal Khirodkar, Devansh Kukreja, Kevin J Liang, Jia-Wei Liu, Sagnik Majumder, Yongsen Mao, Miguel Martin, Effrosyni Mavroudi, Tushar Nagarajan, Francesco Ragusa, Santhosh Kumar Ramakrishnan, Luigi Seminara, Arjun Somayazulu, Yale Song, Shan Su, Zihui Xue, Edward Zhang, Jinxu Zhang, Angela Castillo, Changan Chen, Xinzhu Fu, Ryosuke Furuta, Cristina Gonzalez, Prince Gupta, Jiabo Hu, Yifei HUANG, Yiming Huang, Weslie Khoo, Anush Kumar, Robert Kuo, Sach Lakhavani, Miao Liu, Mi Luo, Zhengyi Luo, Brighid Meredith, Austin Miller, Oluwatumininu Oguntola, Xiaqing Pan, Penny Peng, Shraman Pramanick, Merey Ramazanova, Fiona Ryan, Wei Shan, Kiran Somasundaram, Chenan Song, Audrey Southerland, Masatoshi Tateno, Huiyu Wang, Yuchen Wang, Takuma Yagi, Mingfei Yan, Xitong Yang, Zecheng Yu, Shengxin Cindy Zha, Chen Zhao, Ziwei Zhao, Zhifan Zhu, Jeff Zhuo, Pablo Arbelaez, Gedas Bertasius, David Crandall, Dima Damen, Jakob Engel, Giovanni Maria Farinella, Antonino Furnari, Bernard Ghanem, Judy Hoffman, C. V. Jawahar, Richard Newcombe, Hyun Soo Park, James M. Rehg, Yoichi Sato, Manolis Savva, Jianbo Shi, Mike Zheng Shou, Michael Wray
We present Ego-Exo4D, a diverse, large-scale multimodal multiview video dataset and benchmark challenge.
2 code implementations • 26 Sep 2023 • Francesco Ragusa, Rosario Leonardi, Michele Mazzamuto, Claudia Bonanno, Rosario Scavo, Antonino Furnari, Giovanni Maria Farinella
ENIGMA-51 is a new egocentric dataset acquired in an industrial scenario by 19 subjects who followed instructions to complete the repair of electrical boards using industrial tools (e. g., electric screwdriver) and equipments (e. g., oscilloscope).
no code implementations • 14 Aug 2023 • Chiara Plizzari, Gabriele Goletto, Antonino Furnari, Siddhant Bansal, Francesco Ragusa, Giovanni Maria Farinella, Dima Damen, Tatiana Tommasi
What will the future be?
no code implementations • 29 Jun 2023 • Antonino Furnari, Giovanni Maria Farinella
We propose a streaming egocentric action evaluation scheme which assumes that predictions are performed online and made available only after the model has processed the current input segment, which depends on its runtime.
1 code implementation • 21 Jun 2023 • Rosario Leonardi, Francesco Ragusa, Antonino Furnari, Giovanni Maria Farinella
In this paper, we tackle the problem of Egocentric Human-Object Interaction (EHOI) detection in an industrial setting.
1 code implementation • 8 Apr 2023 • Francesco Ragusa, Giovanni Maria Farinella, Antonino Furnari
Anticipation problem has been studied considering different aspects such as predicting humans' locations, predicting hands and objects trajectories, and forecasting actions and human-object interactions.
1 code implementation • Computer Vision and Image Understanding (CVIU) 2022 • Giovanni Pasqualino, Antonino Furnari, Giovanni Maria Farinella
Object detection algorithms allow to enable many interesting applications which can be implemented in different devices, such as smartphones and wearable devices.
no code implementations • 27 Sep 2022 • Matteo Dunnhofer, Antonino Furnari, Giovanni Maria Farinella, Christian Micheloni
Despite a few previous attempts to exploit trackers in the FPV domain, a methodical analysis of the performance of state-of-the-art trackers is still missing.
no code implementations • 19 Sep 2022 • Francesco Ragusa, Antonino Furnari, Giovanni Maria Farinella
To encourage research in this field, we present MECCANO, a multimodal dataset of egocentric videos to study humans behavior understanding in industrial-like settings.
no code implementations • 14 Apr 2022 • Camillo Quattrocchi, Daniele Di Mauro, Antonino Furnari, Giovanni Maria Farinella
Motivated by this observation, we propose a pipeline which allows to generate synthetic images from 3D models of real environments and real objects.
no code implementations • 14 Apr 2022 • Michele Mazzamuto, Francesco Ragusa, Antonino Furnari, Giovanni Signorello, Giovanni Maria Farinella
Since labeling large amounts of data to train a standard object detector is expensive in terms of costs and time, we propose a weakly supervised version of the task which leans only on gaze data and a frame-level label indicating the class of the attended object.
1 code implementation • 14 Apr 2022 • Rosario Leonardi, Francesco Ragusa, Antonino Furnari, Giovanni Maria Farinella
We consider the problem of detecting Egocentric HumanObject Interactions (EHOIs) in industrial contexts.
no code implementations • 8 Feb 2022 • Ivan Rodin, Antonino Furnari, Dimitrios Mavroeidis, Giovanni Maria Farinella
Experiments show that the performance of current models designed for trimmed action anticipation is very limited and more research on this task is required.
1 code implementation • 2 Feb 2022 • Marco Rosano, Antonino Furnari, Luigi Gulino, Corrado Santoro, Giovanni Maria Farinella
All the proposed navigation models have been trained with the Habitat simulator on a synthetic office environment and have been tested on the same real-world environment using a real robotic platform.
8 code implementations • CVPR 2022 • Kristen Grauman, Andrew Westbury, Eugene Byrne, Zachary Chavis, Antonino Furnari, Rohit Girdhar, Jackson Hamburger, Hao Jiang, Miao Liu, Xingyu Liu, Miguel Martin, Tushar Nagarajan, Ilija Radosavovic, Santhosh Kumar Ramakrishnan, Fiona Ryan, Jayant Sharma, Michael Wray, Mengmeng Xu, Eric Zhongcong Xu, Chen Zhao, Siddhant Bansal, Dhruv Batra, Vincent Cartillier, Sean Crane, Tien Do, Morrie Doulaty, Akshay Erapalli, Christoph Feichtenhofer, Adriano Fragomeni, Qichen Fu, Abrham Gebreselasie, Cristina Gonzalez, James Hillis, Xuhua Huang, Yifei HUANG, Wenqi Jia, Weslie Khoo, Jachym Kolar, Satwik Kottur, Anurag Kumar, Federico Landini, Chao Li, Yanghao Li, Zhenqiang Li, Karttikeya Mangalam, Raghava Modhugu, Jonathan Munro, Tullie Murrell, Takumi Nishiyasu, Will Price, Paola Ruiz Puentes, Merey Ramazanova, Leda Sari, Kiran Somasundaram, Audrey Southerland, Yusuke Sugano, Ruijie Tao, Minh Vo, Yuchen Wang, Xindi Wu, Takuma Yagi, Ziwei Zhao, Yunyi Zhu, Pablo Arbelaez, David Crandall, Dima Damen, Giovanni Maria Farinella, Christian Fuegen, Bernard Ghanem, Vamsi Krishna Ithapu, C. V. Jawahar, Hanbyul Joo, Kris Kitani, Haizhou Li, Richard Newcombe, Aude Oliva, Hyun Soo Park, James M. Rehg, Yoichi Sato, Jianbo Shi, Mike Zheng Shou, Antonio Torralba, Lorenzo Torresani, Mingfei Yan, Jitendra Malik
We introduce Ego4D, a massive-scale egocentric video dataset and benchmark suite.
no code implementations • 11 Oct 2021 • Antonino Furnari, Giovanni Maria Farinella
In contrast, in this paper, we propose a "streaming" egocentric action anticipation evaluation protocol which explicitly considers model runtime for performance assessment, assuming that predictions will be available only after the current video segment is processed, which depends on the processing time of a method.
no code implementations • 31 Aug 2021 • Matteo Dunnhofer, Antonino Furnari, Giovanni Maria Farinella, Christian Micheloni
Our study extensively analyses the performance of recent visual trackers and baseline FPV trackers with respect to different aspects and considering a new performance measure.
no code implementations • 28 Jul 2021 • Ivan Rodin, Antonino Furnari, Dimitrios Mavroedis, Giovanni Maria Farinella
Egocentric videos can bring a lot of information about how humans perceive the world and interact with the environment, which can be beneficial for the analysis of human behaviour.
no code implementations • 22 Jun 2021 • Ronja Möller, Antonino Furnari, Sebastiano Battiato, Aki Härmä, Giovanni Maria Farinella
This paper is concerned with the navigation aspect of a socially-compliant robot and provides a survey of existing solutions for the relevant areas of research as well as an outlook on possible future directions.
no code implementations • 24 Nov 2020 • Matteo Dunnhofer, Antonino Furnari, Giovanni Maria Farinella, Christian Micheloni
Despite a few previous attempts to exploit trackers in FPV applications, a methodical analysis of the performance of state-of-the-art visual trackers in this domain is still missing.
1 code implementation • 26 Oct 2020 • Marco Rosano, Antonino Furnari, Luigi Gulino, Giovanni Maria Farinella
Visual navigation models based on deep learning can learn effective policies when trained on large amounts of visual observations through reinforcement learning.
1 code implementation • 12 Oct 2020 • Francesco Ragusa, Antonino Furnari, Salvatore Livatino, Giovanni Maria Farinella
To fill this gap, we introduce MECCANO, the first dataset of egocentric videos to study human-object interactions in industrial-like settings.
Ranked #1 on Action Recognition on MECCANO
1 code implementation • 4 Aug 2020 • Giovanni Pasqualino, Antonino Furnari, Giovanni Signorello, Giovanni Maria Farinella
To address this problem, we created a new dataset containing both synthetic and real images of 16 different artworks.
Ranked #1 on Unsupervised Domain Adaptation on UDA-CH
7 code implementations • 23 Jun 2020 • Dima Damen, Hazel Doughty, Giovanni Maria Farinella, Antonino Furnari, Evangelos Kazakos, Jian Ma, Davide Moltisanti, Jonathan Munro, Toby Perrett, Will Price, Michael Wray
This paper introduces the pipeline to extend the largest dataset in egocentric vision, EPIC-KITCHENS.
Ranked #8 on Action Anticipation on EPIC-KITCHENS-100
no code implementations • 18 Jun 2020 • Daniele Di Mauro, Antonino Furnari, Giuseppe Patanè, Sebastiano Battiato, Giovanni Maria Farinella
We formalize this problem as a domain adaptation task and introduce a novel dataset of urban scenes with the related semantic labels.
2 code implementations • 4 May 2020 • Antonino Furnari, Giovanni Maria Farinella
The experiments show that the proposed architecture is state-of-the-art in the domain of egocentric videos, achieving top performances in the 2019 EPIC-Kitchens egocentric action anticipation challenge.
Ranked #7 on Action Anticipation on EPIC-KITCHENS-100 (test)
2 code implementations • 29 Apr 2020 • Dima Damen, Hazel Doughty, Giovanni Maria Farinella, Sanja Fidler, Antonino Furnari, Evangelos Kazakos, Davide Moltisanti, Jonathan Munro, Toby Perrett, Will Price, Michael Wray
Our dataset features 55 hours of video consisting of 11. 5M frames, which we densely labelled for a total of 39. 6K action segments and 454. 2K object bounding boxes.
no code implementations • 16 Apr 2020 • Guglielmo Camporese, Pasquale Coscia, Antonino Furnari, Giovanni Maria Farinella, Lamberto Ballan
Since multiple actions may equally occur in the future, we treat action anticipation as a multi-label problem with missing labels extending the concept of label smoothing.
no code implementations • 3 Feb 2020 • Francesco Ragusa, Antonino Furnari, Sebastiano Battiato, Giovanni Signorello, Giovanni Maria Farinella
Equipping visitors of a cultural site with a wearable device allows to easily collect information about their preferences which can be exploited to improve the fruition of cultural goods with augmented reality.
2 code implementations • ICCV 2019 • Antonino Furnari, Giovanni Maria Farinella
Our method is ranked first in the public leaderboard of the EPIC-Kitchens egocentric action anticipation challenge 2019.
Ranked #3 on Egocentric Activity Recognition on EPIC-KITCHENS-55
no code implementations • 10 Apr 2019 • Francesco Ragusa, Antonino Furnari, Sebastiano Battiato, Giovanni Signorello, Giovanni Maria Farinella
We consider the problem of localizing visitors in a cultural site from egocentric (first person) images.
no code implementations • 10 Apr 2019 • Antonino Furnari, Sebastiano Battiato, Kristen Grauman, Giovanni Maria Farinella
Although First Person Vision systems can sense the environment from the user's perspective, they are generally unable to predict his intentions and goals.
2 code implementations • ECCV 2018 • Dima Damen, Hazel Doughty, Giovanni Maria Farinella, Sanja Fidler, Antonino Furnari, Evangelos Kazakos, Davide Moltisanti, Jonathan Munro, Toby Perrett, Will Price, Michael Wray
First-person vision is gaining interest as it offers a unique viewpoint on people's interaction with objects, their attention, and even intention.