3D Action Recognition
34 papers with code • 3 benchmarks • 14 datasets
Image: Rahmani et al
Libraries
Use these libraries to find 3D Action Recognition models and implementationsDatasets
Subtasks
Most implemented papers
Recognizing Involuntary Actions from 3D Skeleton Data Using Body States
This method introduces the definition of body states and then every action is modeled as a sequence of these states.
Fisherposes for Human Action Recognition Using Kinect Sensor Data
The proposed regularized Mahalanobis distance metric is used in order to recognize both the involuntary and highly made-up actions at the same time.
Temporal Transformer Networks: Joint Learning of Invariant and Discriminative Time Warping
We call this a temporal transformer network (TTN).
SkeleMotion: A New Representation of Skeleton Joint Sequences Based on Motion Information for 3D Action Recognition
Due to the availability of large-scale skeleton datasets, 3D human action recognition has recently called the attention of computer vision community.
Skeleton Image Representation for 3D Action Recognition based on Tree Structure and Reference Joints
The proposed representation has the advantage of combining the use of reference joints and a tree structure skeleton.
A System for Real-Time Interactive Analysis of Deep Learning Training
To achieve this, we model various exploratory inspection and diagnostic tasks for deep learning training processes as specifications for streams using a map-reduce paradigm with which many data scientists are already familiar.
3DV: 3D Dynamic Voxel for Action Recognition in Depth Video
Each available 3DV voxel intrinsically involves 3D spatial and motion feature jointly.
Roweisposes, Including Eigenposes, Supervised Eigenposes, and Fisherposes, for 3D Action Recognition
Although various methods have been proposed for 3D action recognition, some of which are basic and some use deep learning, the need of basic methods based on generalized eigenvalue problem is sensed for action recognition.
BABEL: Bodies, Action and Behavior with English Labels
To address this, we present BABEL, a large dataset with language labels describing the actions being performed in mocap sequences.
Point 4D Transformer Networks for Spatio-Temporal Modeling in Point Cloud Videos
To capture the dynamics in point cloud videos, point tracking is usually employed.