Search Results for author: Shengfeng He

Found 63 papers, 31 papers with code

Drag Your Noise: Interactive Point-based Editing via Diffusion Semantic Propagation

1 code implementation1 Apr 2024 Haofeng Liu, Chenshu Xu, Yifei Yang, Lihua Zeng, Shengfeng He

Point-based interactive editing serves as an essential tool to complement the controllability of existing generative models.

Denoising

Learning with Unreliability: Fast Few-shot Voxel Radiance Fields with Relative Geometric Consistency

no code implementations26 Mar 2024 YingJie Xu, Bangzhen Liu, Hao Tang, Bailin Deng, Shengfeng He

We propose a voxel-based optimization framework, ReVoRF, for few-shot radiance fields that strategically address the unreliability in pseudo novel view synthesis.

Novel View Synthesis

Rethinking Multi-view Representation Learning via Distilled Disentangling

1 code implementation16 Mar 2024 Guanzhou Ke, Bo wang, Xiaoli Wang, Shengfeng He

To this end, we propose an innovative framework for multi-view representation learning, which incorporates a technique we term 'distilled disentangling'.

Representation Learning

DiTMoS: Delving into Diverse Tiny-Model Selection on Microcontrollers

1 code implementation14 Mar 2024 Xiao Ma, Shengfeng He, Hezhe Qiao, Dong Ma

Enabling efficient and accurate deep neural network (DNN) inference on microcontrollers is non-trivial due to the constrained on-chip resources.

Emotion Recognition Human Activity Recognition +1

Human Video Translation via Query Warping

no code implementations19 Feb 2024 Haiming Zhu, Yangyang Xu, Shengfeng He

In this paper, we present QueryWarp, a novel framework for temporally coherent human motion video translation.

Denoising Translation +1

DiffusionMat: Alpha Matting as Sequential Refinement Learning

no code implementations22 Nov 2023 Yangyang Xu, Shengfeng He, Wenqi Shao, Kwan-Yee K. Wong, Yu Qiao, Ping Luo

In this paper, we introduce DiffusionMat, a novel image matting framework that employs a diffusion model for the transition from coarse to refined alpha mattes.

Denoising Image Matting

One-for-All: Towards Universal Domain Translation with a Single StyleGAN

no code implementations22 Oct 2023 Yong Du, Jiahui Zhan, Shengfeng He, Xinzhe Li, Junyu Dong, Sheng Chen, Ming-Hsuan Yang

In this paper, we propose a novel translation model, UniTranslator, for transforming representations between visually distinct domains under conditions of limited training data and significant visual differences.

Translation

Delving into Multimodal Prompting for Fine-grained Visual Classification

no code implementations16 Sep 2023 Xin Jiang, Hao Tang, Junyao Gao, Xiaoyu Du, Shengfeng He, Zechao Li

In this paper, we aim to fully exploit the capabilities of cross-modal description to tackle FGVC tasks and propose a novel multimodal prompting solution, denoted as MP-FGVC, based on the contrastive language-image pertaining (CLIP) model.

Classification Fine-Grained Image Classification

NPF-200: A Multi-Modal Eye Fixation Dataset and Method for Non-Photorealistic Videos

1 code implementation23 Aug 2023 Ziyu Yang, Sucheng Ren, Zongwei Wu, Nanxuan Zhao, Junle Wang, Jing Qin, Shengfeng He

Non-photorealistic videos are in demand with the wave of the metaverse, but lack of sufficient research studies.

Saliency Detection

RIGID: Recurrent GAN Inversion and Editing of Real Face Videos

no code implementations ICCV 2023 Yangyang Xu, Shengfeng He, Kwan-Yee K. Wong, Ping Luo

In this paper, we propose a unified recurrent framework, named \textbf{R}ecurrent v\textbf{I}deo \textbf{G}AN \textbf{I}nversion and e\textbf{D}iting (RIGID), to explicitly and simultaneously enforce temporally coherent GAN inversion and facial editing of real videos.

Attribute Facial Editing +1

Disentangling Multi-view Representations Beyond Inductive Bias

1 code implementation3 Aug 2023 Guanzhou Ke, Yang Yu, Guoqing Chao, Xiaoli Wang, Chenyang Xu, Shengfeng He

In this paper, we propose a novel multi-view representation disentangling method that aims to go beyond inductive biases, ensuring both interpretability and generalizability of the resulting representations.

Clustering Inductive Bias +2

Single-View View Synthesis with Self-Rectified Pseudo-Stereo

no code implementations19 Apr 2023 Yang Zhou, Hanjie Wu, Wenxi Liu, Zheng Xiong, Jing Qin, Shengfeng He

In this way, the challenging novel view synthesis process is decoupled into two simpler problems of stereo synthesis and 3D reconstruction.

3D Reconstruction Novel View Synthesis

SCANet: Self-Paced Semi-Curricular Attention Network for Non-Homogeneous Image Dehazing

1 code implementation17 Apr 2023 Yu Guo, Yuan Gao, Ryan Wen Liu, Yuxu Lu, Jingxiang Qu, Shengfeng He, Wenqi Ren

The presence of non-homogeneous haze can cause scene blurring, color distortion, low contrast, and other degradations that obscure texture details.

Image Dehazing

Curricular Contrastive Regularization for Physics-aware Single Image Dehazing

1 code implementation CVPR 2023 Yu Zheng, Jiahui Zhan, Shengfeng He, Junyu Dong, Yong Du

In this paper, we propose a novel curricular contrastive regularization targeted at a consensual contrastive space as opposed to a non-consensual one.

Image Dehazing Single Image Dehazing

Towards a Smaller Student: Capacity Dynamic Distillation for Efficient Image Retrieval

no code implementations CVPR 2023 Yi Xie, Huaidong Zhang, Xuemiao Xu, Jianqing Zhu, Shengfeng He

Specifically, the employed student model is initially a heavy model to fruitfully learn distilled knowledge in the early training epochs, and the student model is gradually compressed during the training.

Image Retrieval Knowledge Distillation +1

Where Is My Spot? Few-Shot Image Generation via Latent Subspace Optimization

1 code implementation CVPR 2023 Chenxi Zheng, Bangzhen Liu, Huaidong Zhang, Xuemiao Xu, Shengfeng He

The rationale behind is that we aim to locate a centroid latent position in a conditional StyleGAN, where the corresponding output image on that centroid can maximize the similarity with the given samples.

Image Generation

Diffuse3D: Wide-Angle 3D Photography via Bilateral Diffusion

1 code implementation ICCV 2023 Yutao Jiang, Yang Zhou, Yuan Liang, Wenxi Liu, Jianbo Jiao, Yuhui Quan, Shengfeng He

To address the above issues, we propose Diffuse3D which employs a pre-trained diffusion model for global synthesis, while amending the model to activate depth-aware inference.

Denoising Novel View Synthesis

Monocular BEV Perception of Road Scenes via Front-to-Top View Projection

no code implementations15 Nov 2022 Wenxi Liu, Qi Li, Weixiang Yang, Jiaxin Cai, Yuanlong Yu, Yuexin Ma, Shengfeng He, Jia Pan

We propose a front-to-top view projection (FTVP) module, which takes the constraint of cycle consistency between views into account and makes full use of their correlation to strengthen the view transformation and scene understanding.

Autonomous Driving Road Segmentation +1

Self-supervised Matting-specific Portrait Enhancement and Generation

1 code implementation13 Aug 2022 Yangyang Xu Zeyang Zhou, Shengfeng He

Particularly, we invert an input portrait into the latent code of StyleGAN, and our aim is to discover whether there is an enhanced version in the latent space which is more compatible with a reference matting model.

Image Matting Specificity

Editing Out-of-domain GAN Inversion via Differential Activations

1 code implementation17 Jul 2022 Haorui Song, Yong Du, Tianyi Xiang, Junyu Dong, Jing Qin, Shengfeng He

Consequently, in the decomposition phase, we further present a GAN prior based deghosting network for separating the final fine edited image from the coarse reconstruction.

Attribute

A Simple Data Mixing Prior for Improving Self-Supervised Learning

1 code implementation CVPR 2022 Sucheng Ren, Huiyu Wang, Zhengqi Gao, Shengfeng He, Alan Yuille, Yuyin Zhou, Cihang Xie

More notably, our SDMP is the first method that successfully leverages data mixing to improve (rather than hurt) the performance of Vision Transformers in the self-supervised setting.

Representation Learning Self-Supervised Learning

Glance to Count: Learning to Rank with Anchors for Weakly-supervised Crowd Counting

no code implementations29 May 2022 Zheng Xiong, Liangyu Chai, Wenxi Liu, Yongtuo Liu, Sucheng Ren, Shengfeng He

To enable training under this new setting, we convert the crowd count regression problem to a ranking potential prediction problem.

Crowd Counting Learning-To-Rank

High-resolution Face Swapping via Latent Semantics Disentanglement

1 code implementation CVPR 2022 Yangyang Xu, Bailin Deng, Junle Wang, Yanqing Jing, Jia Pan, Shengfeng He

Although previous research can leverage generative priors to produce high-resolution results, their quality can suffer from the entangled semantics of the latent space.

Disentanglement Face Swapping +2

Faithful Extreme Rescaling via Generative Prior Reciprocated Invertible Representations

1 code implementation CVPR 2022 Zhixuan Zhong, Liangyu Chai, Yang Zhou, Bailin Deng, Jia Pan, Shengfeng He

This paper presents a Generative prior ReciprocAted Invertible rescaling Network (GRAIN) for generating faithful high-resolution (HR) images from low-resolution (LR) invertible images with an extreme upscaling factor (64x).

Shunted Self-Attention via Multi-Scale Token Aggregation

1 code implementation CVPR 2022 Sucheng Ren, Daquan Zhou, Shengfeng He, Jiashi Feng, Xinchao Wang

This novel merging scheme enables the self-attention to learn relationships between objects with different sizes and simultaneously reduces the token numbers and the computational cost.

Ultra-high Resolution Image Segmentation via Locality-aware Context Fusion and Alternating Local Enhancement

1 code implementation ICCV 2021 Wenxi Liu, Qi Li, Xindai Lin, Weixiang Yang, Shengfeng He, Yuanlong Yu

In particular, we introduce a novel locality-aware context fusion based segmentation model to process local patches, where the relevance between local patch and its various contexts are jointly and complementarily utilized to handle the semantic regions with large variations.

Image Segmentation Land Cover Classification +2

Fine-grained Domain Adaptive Crowd Counting via Point-derived Segmentation

no code implementations6 Aug 2021 Yongtuo Liu, Dan Xu, Sucheng Ren, Hanjie Wu, Hongmin Cai, Shengfeng He

To this end, we propose to untangle \emph{domain-invariant} crowd and \emph{domain-specific} background from crowd images and design a fine-grained domain adaption method for crowd counting.

Crowd Counting Domain Adaptation +1

Reducing Spatial Labeling Redundancy for Semi-supervised Crowd Counting

no code implementations6 Aug 2021 Yongtuo Liu, Sucheng Ren, Liangyu Chai, Hanjie Wu, Jing Qin, Dan Xu, Shengfeng He

In this way, we can transfer the original spatial labeling redundancy caused by individual similarities to effective supervision signals on the unlabeled regions.

Crowd Counting

Unifying Global-Local Representations in Salient Object Detection with Transformer

1 code implementation5 Aug 2021 Sucheng Ren, Qiang Wen, Nanxuan Zhao, Guoqiang Han, Shengfeng He

In this paper, we introduce a new attention-based encoder, vision transformer, into salient object detection to ensure the globalization of the representations from shallow to deep layers.

object-detection Object Detection +1

From Continuity to Editability: Inverting GANs with Consecutive Images

2 code implementations ICCV 2021 Yangyang Xu, Yong Du, Wenpeng Xiao, Xuemiao Xu, Shengfeng He

This inborn property is used for two unique purposes: 1) regularizing the joint inversion process, such that each of the inverted code is semantically accessible from one of the other and fastened in a editable domain; 2) enforcing inter-image coherence, such that the fidelity of each inverted code can be maximized with the complement of other images.

Co-advise: Cross Inductive Bias Distillation

no code implementations CVPR 2022 Sucheng Ren, Zhengqi Gao, Tianyu Hua, Zihui Xue, Yonglong Tian, Shengfeng He, Hang Zhao

Transformers recently are adapted from the community of natural language processing as a promising substitute of convolution-based neural networks for visual learning tasks.

Inductive Bias

Reciprocal Transformations for Unsupervised Video Object Segmentation

1 code implementation CVPR 2021 Sucheng Ren, Wenxi Liu, Yongtuo Liu, Haoxin Chen, Guoqiang Han, Shengfeng He

Additionally, to exclude the information of the moving background objects from motion features, our transformation module enables to reciprocally transform the appearance features to enhance the motion features, so as to focus on the moving objects with salient appearance while removing the co-moving outliers.

Object Optical Flow Estimation +3

Learning From the Master: Distilling Cross-Modal Advanced Knowledge for Lip Reading

no code implementations CVPR 2021 Sucheng Ren, Yong Du, Jianming Lv, Guoqiang Han, Shengfeng He

To these ends, we introduce a trainable "master" network which ingests both audio signals and silent lip videos instead of a pretrained teacher.

Lip Reading Sentence +2

Discovering Interpretable Latent Space Directions of GANs Beyond Binary Attributes

1 code implementation CVPR 2021 Huiting Yang, Liangyu Chai, Qiang Wen, Shuang Zhao, Zixun Sun, Shengfeng He

In this way, arbitrary attributes can be edited by collecting positive data only, and the proposed method learns a controllable representation enabling manipulation of non-binary attributes like anime styles and facial characteristics.

Attribute

Spatially-Invariant Style-Codes Controlled Makeup Transfer

1 code implementation CVPR 2021 Han Deng, Chu Han, Hongmin Cai, Guoqiang Han, Shengfeng He

In this paper, we take a different perspective to break down the makeup transfer problem into a two-step extraction-assignment process.

Smart Scribbles for Image Mating

no code implementations31 Mar 2021 Xin Yang, Yu Qiao, Shaozhe Chen, Shengfeng He, BaoCai Yin, Qiang Zhang, Xiaopeng Wei, Rynson W. H. Lau

Image matting is an ill-posed problem that usually requires additional user input, such as trimaps or scribbles.

Image Matting

Self-supervised Video Representation Learning by Uncovering Spatio-temporal Statistics

2 code implementations31 Aug 2020 Jiangliu Wang, Jianbo Jiao, Linchao Bao, Shengfeng He, Wei Liu, Yun-hui Liu

Specifically, given an unlabeled video clip, we compute a series of spatio-temporal statistical summaries, such as the spatial location and dominant direction of the largest motion, the spatial location and dominant color of the largest color diversity along the temporal axis, etc.

Action Recognition Representation Learning +3

TENet: Triple Excitation Network for Video Salient Object Detection

no code implementations ECCV 2020 Sucheng Ren, Chu Han, Xin Yang, Guoqiang Han, Shengfeng He

In this paper, we propose a simple yet effective approach, named Triple Excitation Network, to reinforce the training of video salient object detection (VSOD) from three aspects, spatial, temporal, and online excitations.

Object object-detection +2

Over-crowdedness Alert! Forecasting the Future Crowd Distribution

no code implementations9 Jun 2020 Yuzhen Niu, Weifeng Shi, Wenxi Liu, Shengfeng He, Jia Pan, Antoni B. Chan

In this paper, we formulate a novel crowd analysis problem, in which we aim to predict the crowd distribution in the near future given sequential frames of a crowd video without any identity annotations.

Context-aware and Scale-insensitive Temporal Repetition Counting

1 code implementation CVPR 2020 Huaidong Zhang, Xuemiao Xu, Guoqiang Han, Shengfeng He

It avoids the heavy computation of exhaustively searching all the cycle lengths in the video, and, instead, it propagates the coarse prediction for further refinement in a hierarchical manner.

regression

Laplacian Denoising Autoencoder

no code implementations30 Mar 2020 Jianbo Jiao, Linchao Bao, Yunchao Wei, Shengfeng He, Honghui Shi, Rynson Lau, Thomas S. Huang

This can be naturally generalized to span multiple scales with a Laplacian pyramid representation of the input data.

Denoising Self-Supervised Learning

Visualizing the Invisible: Occluded Vehicle Segmentation and Recovery

no code implementations ICCV 2019 Xiaosheng Yan, Yuanlong Yu, Feigege Wang, Wenxi Liu, Shengfeng He, Jia Pan

We conduct comparison experiments on this dataset and demonstrate that our model outperforms the state-of-the-art in tasks of recovering segmentation mask and appearance for occluded vehicles.

Segmentation

Active Matting

no code implementations NeurIPS 2018 Xin Yang, Ke Xu, Shaozhe Chen, Shengfeng He, Baocai Yin Yin, Rynson Lau

Our aim is to discover the most informative sequence of regions for user input in order to produce a good alpha matte with minimum labeling efforts.

Image Matting

Joint Face Hallucination and Deblurring via Structure Generation and Detail Enhancement

no code implementations22 Nov 2018 Yibing Song, Jiawei Zhang, Lijun Gong, Shengfeng He, Linchao Bao, Jinshan Pan, Qingxiong Yang, Ming-Hsuan Yang

We first propose a facial component guided deep Convolutional Neural Network (CNN) to restore a coarse face image, which is denoted as the base image where the facial component is automatically generated from the input face image.

Deblurring Face Hallucination +2

Deformable Object Tracking with Gated Fusion

no code implementations27 Sep 2018 Wenxi Liu, Yibing Song, Dengsheng Chen, Shengfeng He, Yuanlong Yu, Tao Yan, Gerhard P. Hancke, Rynson W. H. Lau

In addition, we also propose a gated fusion scheme to control how the variations captured by the deformable convolution affect the original appearance.

Object Object Tracking

SINet: A Scale-insensitive Convolutional Neural Network for Fast Vehicle Detection

no code implementations2 Apr 2018 Xiaowei Hu, Xuemiao Xu, Yongjie Xiao, Hao Chen, Shengfeng He, Jing Qin, Pheng-Ann Heng

Based on these findings, we present a scale-insensitive convolutional neural network (SINet) for fast detecting vehicles with a large variance of scales.

Fast Vehicle Detection object-detection +1

Egocentric Hand Detection Via Dynamic Region Growing

no code implementations10 Nov 2017 Shao Huang, Weiqiang Wang, Shengfeng He, Rynson W. H. Lau

Egocentric videos, which mainly record the activities carried out by the users of the wearable cameras, have drawn much research attentions in recent years.

Action Recognition Gesture Recognition +3

Delving Into Salient Object Subitizing and Detection

no code implementations ICCV 2017 Shengfeng He, Jianbo Jiao, Xiaodan Zhang, Guoqiang Han, Rynson W. H. Lau

Experiments show that the proposed multi-task network outperforms existing multi-task architectures, and the auxiliary subitizing network provides strong guidance to salient object detection by reducing false positives and producing coherent saliency maps.

Object object-detection +2

Stylizing Face Images via Multiple Exemplars

no code implementations28 Aug 2017 Yibing Song, Linchao Bao, Shengfeng He, Qingxiong Yang, Ming-Hsuan Yang

We address the problem of transferring the style of a headshot photo to face images.

RGBD Salient Object Detection via Deep Fusion

no code implementations12 Jul 2016 Liangqiong Qu, Shengfeng He, Jiawei Zhang, Jiandong Tian, Yandong Tang, Qingxiong Yang

Numerous efforts have been made to design different low level saliency cues for the RGBD saliency detection, such as color or depth contrast features, background and color compactness priors.

Object object-detection +4

Real-Time Salient Object Detection With a Minimum Spanning Tree

1 code implementation CVPR 2016 Wei-Chih Tu, Shengfeng He, Qingxiong Yang, Shao-Yi Chien

In this paper, we present a real-time salient object detection system based on the minimum spanning tree.

Ranked #5 on Video Salient Object Detection on MCL (using extra training data)

Object object-detection +2

Exemplar-Driven Top-Down Saliency Detection via Deep Association

no code implementations CVPR 2016 Shengfeng He, Rynson W. H. Lau, Qingxiong Yang

To address it, we design a two-stage deep model to learn the intra-class association between the exemplars and query objects.

Object Saliency Detection

Oriented Object Proposals

no code implementations ICCV 2015 Shengfeng He, Rynson W. H. Lau

In this paper, we propose a new approach to generate oriented object proposals (OOPs) to reduce the detection error caused by various orientations of the object.

Object

Visual Tracking via Locality Sensitive Histograms

no code implementations CVPR 2013 Shengfeng He, Qingxiong Yang, Rynson W. H. Lau, Jiang Wang, Ming-Hsuan Yang

A robust tracking framework based on the locality sensitive histograms is proposed, which consists of two main components: a new feature for tracking that is robust to illumination changes and a novel multi-region tracking algorithm that runs in realtime even with hundreds of regions.

Visual Tracking

Cannot find the paper you are looking for? You can Submit a new open access paper.