Search Results for author: Xiaojuan Wang

Found 11 papers, 3 papers with code

Jump Cut Smoothing for Talking Heads

no code implementations9 Jan 2024 Xiaojuan Wang, Taesung Park, Yang Zhou, Eli Shechtman, Richard Zhang

We leverage the appearance of the subject from the other source frames in the video, fusing it with a mid-level representation driven by DensePose keypoints and face landmarks.

Generative Powers of Ten

no code implementations4 Dec 2023 Xiaojuan Wang, Janne Kontkanen, Brian Curless, Steve Seitz, Ira Kemelmacher, Ben Mildenhall, Pratul Srinivasan, Dor Verbin, Aleksander Holynski

We present a method that uses a text-to-image model to generate consistent content across multiple image scales, enabling extreme semantic zooms into a scene, e. g., ranging from a wide-angle landscape view of a forest to a macro shot of an insect sitting on one of the tree branches.

Image Super-Resolution

QueryPose: Sparse Multi-Person Pose Regression via Spatial-Aware Part-Level Query

3 code implementations15 Dec 2022 Yabo Xiao, Kai Su, Xiaojuan Wang, Dongdong Yu, Lei Jin, Mingshu He, Zehuan Yuan

The existing end-to-end methods rely on dense representations to preserve the spatial detail and structure for precise keypoint localization.

regression

AdaptivePose++: A Powerful Single-Stage Network for Multi-Person Pose Regression

1 code implementation8 Oct 2022 Yabo Xiao, Xiaojuan Wang, Dongdong Yu, Kai Su, Lei Jin, Mei Song, Shuicheng Yan, Jian Zhao

With the proposed body representation, we further deliver a compact single-stage multi-person pose regression network, termed as AdaptivePose.

3D Multi-Person Pose Estimation Human Detection +2

Learning Quality-aware Representation for Multi-person Pose Regression

no code implementations4 Jan 2022 Yabo Xiao, Dongdong Yu, Xiaojuan Wang, Lei Jin, Guoli Wang, Qian Zhang

Off-the-shelf single-stage multi-person pose regression methods generally leverage the instance score (i. e., confidence of the instance localization) to indicate the pose quality for selecting the pose candidates.

regression

Single-Stage Is Enough: Multi-Person Absolute 3D Pose Estimation

no code implementations CVPR 2022 Lei Jin, Chenyang Xu, Xiaojuan Wang, Yabo Xiao, Yandong Guo, Xuecheng Nie, Jian Zhao

The existing multi-person absolute 3D pose estimation methods are mainly based on two-stage paradigm, i. e., top-down or bottom-up, leading to redundant pipelines with high computation cost.

3D Pose Estimation Depth Estimation +1

AdaptivePose: Human Parts as Adaptive Points

1 code implementation27 Dec 2021 Yabo Xiao, Xiaojuan Wang, Dongdong Yu, Guoli Wang, Qian Zhang, Mingshu He

Multi-person pose estimation methods generally follow top-down and bottom-up paradigms, both of which can be considered as two-stage approaches thus leading to the high computation cost and low efficiency.

Multi-Person Pose Estimation

The 2nd Anti-UAV Workshop & Challenge: Methods and Results

no code implementations23 Aug 2021 Jian Zhao, Gang Wang, Jianan Li, Lei Jin, Nana Fan, Min Wang, Xiaojuan Wang, Ting Yong, Yafeng Deng, Yandong Guo, Shiming Ge, Guodong Guo

The 2nd Anti-UAV Workshop \& Challenge aims to encourage research in developing novel and accurate methods for multi-scale object tracking.

Object Tracking

SPCNet:Spatial Preserve and Content-aware Network for Human Pose Estimation

no code implementations13 Apr 2020 Yabo Xiao, Dongdong Yu, Xiaojuan Wang, Tianqi Lv, Yiqi Fan, Lingrui Wu

To alleviate these issues, we propose a novel Spatial Preserve and Content-aware Network(SPCNet), which includes two effective modules: Dilated Hourglass Module(DHM) and Selective Information Module(SIM).

Pose Estimation

Multi-Scale Learning for Low-Resolution Person Re-Identification

no code implementations ICCV 2015 Xiang Li, Wei-Shi Zheng, Xiaojuan Wang, Tao Xiang, Shaogang Gong

In real world person re-identification (re-id), images of people captured at very different resolutions from different locations need be matched.

Person Re-Identification

Cannot find the paper you are looking for? You can Submit a new open access paper.