Long Term Action Anticipation

6 papers with code • 1 benchmarks • 1 datasets

This task has no description! Would you like to contribute one?

Datasets


Most implemented papers

Video + CLIP Baseline for Ego4D Long-term Action Anticipation

srijandas07/clip_baseline_lta_ego4d 1 Jul 2022

The CLIP embedding provides fine-grained understanding of objects relevant for an action whereas the slowfast network is responsible for modeling temporal information within a video clip of few frames.

Intention-Conditioned Long-Term Human Egocentric Action Forecasting

evm7/ego4dlta-icvae 25 Jul 2022

Our framework first extracts two level of human information over the N observed videos human actions through a Hierarchical Multi-task MLP Mixer (H3M).

Learning State-Aware Visual Representations from Audible Interactions

HimangiM/RepLAI 27 Sep 2022

However, learning representations from videos can be challenging.

Rethinking Learning Approaches for Long-Term Action Anticipation

nmegha2601/anticipatr 20 Oct 2022

Action anticipation involves predicting future actions having observed the initial portion of a video.

Palm: Predicting Actions through Language Models @ Ego4D Long-Term Action Anticipation Challenge 2023

dandoge/palm 28 Jun 2023

We present Palm, a solution to the Long-Term Action Anticipation (LTA) task utilizing vision-language and large language models.

Object-centric Video Representation for Long-term Action Anticipation

brown-palm/ObjectPrompt 31 Oct 2023

To recognize and predict human-object interactions, we use a Transformer-based neural architecture which allows the "retrieval" of relevant objects for action anticipation at various time scales.