|Trend||Dataset||Best Method||Paper title||Paper||Code||Compare|
Detecting activities in untrimmed videos is an important but challenging task.
Temporal action proposal generation is an important yet challenging problem, since temporal proposals with rich action content are indispensable for analysing real-world videos with long duration and high proportion irrelevant content.
#2 best model for Temporal Action Proposal Generation on ActivityNet-1.3
This thesis explore different approaches using Convolutional and Recurrent Neural Networks to classify and temporally localize activities on videos, furthermore an implementation to achieve it has been proposed.
In this paper, we introduce the concept of learning latent super-events from activity videos, and present how it benefits activity detection in continuous videos.
#2 best model for Action Detection on Charades (using extra training data)
While observing complex events with multiple actors, humans do not assess each actor separately, but infer from the context.
Diffusions effectively interact two aspects of information, i. e., localized and holistic, for more powerful way of representation learning.
A test video is processed by forming correspondences between its clips and the clips of reference videos with known semantics, following which, reference semantics can be transferred to the test video.