Motion Captioning
6 papers with code • 2 benchmarks • 2 datasets
Generating textual description for human motion.
Most implemented papers
MotionGPT: Human Motion as a Foreign Language
Building upon this "motion vocabulary", we perform language modeling on both motion and text in a unified manner, treating human motion as a specific language.
TM2T: Stochastic and Tokenized Modeling for the Reciprocal Generation of 3D Human Motions and Texts
Our approach is flexible, could be used for both text2motion and motion2text tasks.
Guided Attention for Interpretable Motion Captioning
Diverse and extensive work has recently been conducted on text-conditioned human motion generation.
Motion2Language, unsupervised learning of synchronized semantic motion segmentation
We find that both contributions to the attention mechanism and the encoder architecture additively improve the quality of generated text (BLEU and semantic equivalence), but also of synchronization.
Motion-Agent: A Conversational Framework for Human Motion Generation with LLMs
This is accomplished by encoding and quantizing motions into discrete tokens that align with the language model's vocabulary.
Transformer with Controlled Attention for Synchronous Motion Captioning
In this paper, we address a challenging task, synchronous motion captioning, that aim to generate a language description synchronized with human motion sequences.