Search Results for author: Shuliang Ning

Found 5 papers, 2 papers with code

PICTURE: PhotorealistIC virtual Try-on from UnconstRained dEsigns

no code implementations7 Dec 2023 Shuliang Ning, Duomin Wang, Yipeng Qin, Zirong Jin, Baoyuan Wang, Xiaoguang Han

Unlike prior arts constrained by specific input types, our method allows flexible specification of style (text or image) and texture (full garment, cropped sections, or texture patches) conditions.

Disentanglement Human Parsing +1

FashionTex: Controllable Virtual Try-on with Text and Texture

1 code implementation8 May 2023 Anran Lin, Nanxuan Zhao, Shuliang Ning, Yuda Qiu, Baoyuan Wang, Xiaoguang Han

Virtual try-on attracts increasing research attention as a promising way for enhancing the user experience for online cloth shopping.

Virtual Try-on

MIMO Is All You Need : A Strong Multi-In-Multi-Out Baseline for Video Prediction

1 code implementation9 Dec 2022 Shuliang Ning, Mengcheng Lan, Yanran Li, Chaofeng Chen, Qian Chen, Xunlai Chen, Xiaoguang Han, Shuguang Cui

The mainstream of the existing approaches for video prediction builds up their models based on a Single-In-Single-Out (SISO) architecture, which takes the current frame as input to predict the next frame in a recursive manner.

Video Prediction

From Single to Multiple: Leveraging Multi-level Prediction Spaces for Video Forecasting

no code implementations21 Jul 2021 Mengcheng Lan, Shuliang Ning, Yanran Li, Qian Chen, Xunlai Chen, Xiaoguang Han, Shuguang Cui

Despite video forecasting has been a widely explored topic in recent years, the mainstream of the existing work still limits their models with a single prediction space but completely neglects the way to leverage their model with multi-prediction spaces.

Video Prediction

Cannot find the paper you are looking for? You can Submit a new open access paper.