Search Results for author: Linze Li

Found 8 papers, 6 papers with code

MegActor-$Σ$: Unlocking Flexible Mixed-Modal Control in Portrait Animation with Diffusion Transformer

2 code implementations27 Aug 2024 Shurong Yang, Huadong Li, Juhao Wu, Minhao Jing, Linze Li, Renhe Ji, Jiajun Liang, Haoqiang Fan, Jin Wang

To address this issue, we introduce MegActor-$\Sigma$: a mixed-modal conditional diffusion transformer (DiT), which can flexibly inject audio and visual modality control signals into portrait animation.

Portrait Animation

Efficient Single Image Super-Resolution with Entropy Attention and Receptive Field Augmentation

no code implementations8 Aug 2024 Xiaole Zhao, Linze Li, Chengxing Xie, XiaoMing Zhang, Ting Jiang, Wenjie Lin, Shuaicheng Liu, Tianrui Li

Transformer-based deep models for single image super-resolution (SISR) have greatly improved the performance of lightweight SISR tasks in recent years.

Image Super-Resolution

Large Kernel Distillation Network for Efficient Single Image Super-Resolution

1 code implementation19 Jul 2024 Chengxing Xie, XiaoMing Zhang, Linze Li, Haiteng Meng, Tianlin Zhang, Tianrui Li, Xiaole Zhao

Efficient and lightweight single-image super-resolution (SISR) has achieved remarkable performance in recent years.

Image Super-Resolution

C2C: Component-to-Composition Learning for Zero-Shot Compositional Action Recognition

1 code implementation8 Jul 2024 Rongchang Li, ZhenHua Feng, Tianyang Xu, Linze Li, Xiao-Jun Wu, Muhammad Awais, Sara Atito, Josef Kittler

For evaluating the task, we construct a new benchmark, Something-composition (Sth-com), based on the widely used Something-Something V2 dataset.

Action Recognition

MegActor: Harness the Power of Raw Video for Vivid Portrait Animation

2 code implementations31 May 2024 Shurong Yang, Huadong Li, Juhao Wu, Minhao Jing, Linze Li, Renhe Ji, Jiajun Liang, Haoqiang Fan

Despite raw driving videos contain richer information on facial expressions than intermediate representations such as landmarks in the field of portrait animation, they are seldom the subject of research.

Portrait Animation Style Transfer +1

TENet: Targetness Entanglement Incorporating with Multi-Scale Pooling and Mutually-Guided Fusion for RGB-E Object Tracking

1 code implementation8 May 2024 Pengcheng Shao, Tianyang Xu, Zhangyong Tang, Linze Li, Xiao-Jun Wu, Josef Kittler

There is currently strong interest in improving visual object tracking by augmenting the RGB modality with the output of a visual event camera that is particularly informative about the scene motion.

Visual Object Tracking

FAAC: Facial Animation Generation with Anchor Frame and Conditional Control for Superior Fidelity and Editability

no code implementations6 Dec 2023 Linze Li, Sunqi Fan, Hengjun Pu, Zhaodong Bing, Yao Tang, Tianzhu Ye, Tong Yang, Liangyu Chen, Jiajun Liang

Our method's efficacy has been validated on multiple representative DreamBooth and LoRA models, delivering substantial improvements over the original outcomes in terms of facial fidelity, text-to-image editability, and video motion.

Face Model Video Generation

Efficient One Pass Self-distillation with Zipf's Label Smoothing

1 code implementation26 Jul 2022 Jiajun Liang, Linze Li, Zhaodong Bing, Borui Zhao, Yao Tang, Bo Lin, Haoqiang Fan

This paper proposes an efficient self-distillation method named Zipf's Label Smoothing (Zipf's LS), which uses the on-the-fly prediction of a network to generate soft supervision that conforms to Zipf distribution without using any contrastive samples or auxiliary parameters.

Cannot find the paper you are looking for? You can Submit a new open access paper.