Search Results for author: Shunlin Lu

Found 10 papers, 3 papers with code

Motion-X++: A Large-Scale Multimodal 3D Whole-body Human Motion Dataset

no code implementations9 Jan 2025 Yuhong Zhang, Jing Lin, Ailing Zeng, Guanlin Wu, Shunlin Lu, Yurong Fu, Yuanhao Cai, Ruimao Zhang, Haoqian Wang, Lei Zhang

To address this issue, we develop a scalable annotation pipeline that can automatically capture 3D whole-body human motion and comprehensive textural labels from RGB videos and build the Motion-X dataset comprising 81. 1K text-motion pairs.

Human Mesh Recovery Motion Generation

ScaMo: Exploring the Scaling Law in Autoregressive Motion Generation Model

no code implementations19 Dec 2024 Shunlin Lu, Jingbo Wang, Zeyu Lu, Ling-Hao Chen, Wenxun Dai, Junting Dong, Zhiyang Dou, Bo Dai, Ruimao Zhang

In this paper, we introduce a scalable motion generation framework that includes the motion tokenizer Motion FSQ-VAE and a text-prefix autoregressive transformer.

Motion Generation

MotionCLR: Motion Generation and Training-free Editing via Understanding Attention Mechanisms

no code implementations24 Oct 2024 Ling-Hao Chen, Wenxun Dai, Xuan Ju, Shunlin Lu, Lei Zhang

Previous motion diffusion models lack explicit modeling of the word-level text-motion correspondence and good explainability, hence restricting their fine-grained editing ability.

Motion Generation

Story3D-Agent: Exploring 3D Storytelling Visualization with Large Language Models

no code implementations21 Aug 2024 Yuzhou Huang, Yiran Qin, Shunlin Lu, Xintao Wang, Rui Huang, Ying Shan, Ruimao Zhang

Traditional visual storytelling is complex, requiring specialized knowledge and substantial resources, yet often constrained by human creativity and creation precision.

Logical Reasoning Motion Synthesis +1

MotionLLM: Understanding Human Behaviors from Human Motions and Videos

1 code implementation30 May 2024 Ling-Hao Chen, Shunlin Lu, Ailing Zeng, Hao Zhang, Benyou Wang, Ruimao Zhang, Lei Zhang

This study delves into the realm of multi-modality (i. e., video and motion modalities) human behavior understanding by leveraging the powerful capabilities of Large Language Models (LLMs).

HumanTOMATO: Text-aligned Whole-body Motion Generation

1 code implementation19 Oct 2023 Shunlin Lu, Ling-Hao Chen, Ailing Zeng, Jing Lin, Ruimao Zhang, Lei Zhang, Heung-Yeung Shum

This work targets a novel text-driven whole-body motion generation task, which takes a given textual description as input and aims at generating high-quality, diverse, and coherent facial expressions, hand gestures, and body motions simultaneously.

Motion Generation Motion Synthesis

Learning to Linearize Deep Neural Networks for Secure and Efficient Private Inference

no code implementations23 Jan 2023 Souvik Kundu, Shunlin Lu, Yuke Zhang, Jacqueline Liu, Peter A. Beerel

For a similar ReLU budget SENet can yield models with ~2. 32% improved classification accuracy, evaluated on CIFAR-100.

Sparse Mixture Once-for-all Adversarial Training for Efficient In-Situ Trade-Off Between Accuracy and Robustness of DNNs

no code implementations27 Dec 2022 Souvik Kundu, Sairam Sundaresan, Sharath Nittur Sridhar, Shunlin Lu, Han Tang, Peter A. Beerel

Existing deep neural networks (DNNs) that achieve state-of-the-art (SOTA) performance on both clean and adversarially-perturbed images rely on either activation or weight conditioned convolution operations.

Image Classification

Cannot find the paper you are looking for? You can Submit a new open access paper.