Search Results for author: Pei Lin

Found 6 papers, 0 papers with code

Dual-Space Optimization: Improved Molecule Sequence Design by Latent Prompt Transformer

no code implementations27 Feb 2024 Deqian Kong, Yuhao Huang, Jianwen Xie, Edouardo Honig, Ming Xu, Shuanghong Xue, Pei Lin, Sanping Zhou, Sheng Zhong, Nanning Zheng, Ying Nian Wu

Designing molecules with desirable properties, such as drug-likeliness and high binding affinities towards protein targets, is a challenging problem.

HandDiffuse: Generative Controllers for Two-Hand Interactions via Diffusion Models

no code implementations8 Dec 2023 Pei Lin, Sihang Xu, Hongdi Yang, Yiran Liu, Xin Chen, Jingya Wang, Jingyi Yu, Lan Xu

We further present a strong baseline method HandDiffuse for the controllable motion generation of interacting hands using various controllers.

Data Augmentation Temporal Sequences

HumanNeRF: Efficiently Generated Human Radiance Field from Sparse Inputs

no code implementations CVPR 2022 Fuqiang Zhao, Wei Yang, Jiakai Zhang, Pei Lin, Yingliang Zhang, Jingyi Yu, Lan Xu

The raw HumanNeRF can already produce reasonable rendering on sparse video inputs of unseen subjects and camera settings.

iButter: Neural Interactive Bullet Time Generator for Human Free-viewpoint Rendering

no code implementations12 Aug 2021 Liao Wang, Ziyu Wang, Pei Lin, Yuheng Jiang, Xin Suo, Minye Wu, Lan Xu, Jingyi Yu

To fill this gap, in this paper we propose a neural interactive bullet-time generator (iButter) for photo-realistic human free-viewpoint rendering from dense RGB streams, which enables flexible and interactive design for human bullet-time visual effects.

Video Generation

Neural Free-Viewpoint Performance Rendering under Complex Human-object Interactions

no code implementations1 Aug 2021 Guoxing Sun, Xin Chen, Yizhang Chen, Anqi Pang, Pei Lin, Yuheng Jiang, Lan Xu, Jingya Wang, Jingyi Yu

In this paper, we propose a neural human performance capture and rendering system to generate both high-quality geometry and photo-realistic texture of both human and objects under challenging interaction scenarios in arbitrary novel views, from only sparse RGB streams.

4D reconstruction Dynamic Reconstruction +5

NeuralHumanFVV: Real-Time Neural Volumetric Human Performance Rendering using RGB Cameras

no code implementations CVPR 2021 Xin Suo, Yuheng Jiang, Pei Lin, Yingliang Zhang, Kaiwen Guo, Minye Wu, Lan Xu

4D reconstruction and rendering of human activities is critical for immersive VR/AR experience. Recent advances still fail to recover fine geometry and texture results with the level of detail present in the input images from sparse multi-view RGB cameras.

4D reconstruction Multi-Task Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.