Search Results for author: Xiaojie Li

Found 8 papers, 4 papers with code

Local Correlation Consistency for Knowledge Distillation

no code implementations ECCV 2020 Xiaojie Li, Jianlong Wu, Hongyu Fang, Yue Liao, Fei Wang, Chen Qian

Sufficient knowledge extraction from the teacher network plays a critical role in the knowledge distillation task to improve the performance of the student network.

Knowledge Distillation

Cross-Modality Domain Adaptation for Freespace Detection: A Simple yet Effective Baseline

no code implementations6 Oct 2022 Yuanbin Wang, Leyan Zhu, Shaofei Huang, Tianrui Hui, Xiaojie Li, Fei Wang, Si Liu

To better bridge the domain gap between source domain (synthetic data) and target domain (real-world data), we also propose a Selective Feature Alignment (SFA) module which only aligns the features of consistent foreground area between the two domains, thus realizing inter-domain intra-modality adaptation.

Autonomous Driving Semantic Segmentation +1

HEAD: HEtero-Assists Distillation for Heterogeneous Object Detectors

1 code implementation12 Jul 2022 Luting Wang, Xiaojie Li, Yue Liao, Zeren Jiang, Jianlong Wu, Fei Wang, Chen Qian, Si Liu

We observe that the core difficulty for heterogeneous KD (hetero-KD) is the significant semantic gap between the backbone features of heterogeneous detectors due to the different optimization manners.

Knowledge Distillation object-detection +2

Face Deblurring Based on Separable Normalization and Adaptive Denormalization

no code implementations18 Dec 2021 Xian Zhang, Hao Zhang, Jiancheng Lv, Xiaojie Li

Face deblurring aims to restore a clear face image from a blurred input image with more explicit structure and facial details.

Deblurring Face Parsing +1

Agree to Disagree: Adaptive Ensemble Knowledge Distillation in Gradient Space

1 code implementation NeurIPS 2020 Shangchen Du, Shan You, Xiaojie Li, Jianlong Wu, Fei Wang, Chen Qian, ChangShui Zhang

In this paper, we examine the diversity of teacher models in the gradient space and regard the ensemble knowledge distillation as a multi-objective optimization problem so that we can determine a better optimization direction for the training of student network.

Knowledge Distillation

Domain Embedded Multi-model Generative Adversarial Networks for Image-based Face Inpainting

no code implementations5 Feb 2020 Xian Zhang, Xin Wang, Bin Kong, Youbing Yin, Qi Song, Siwei Lyu, Jiancheng Lv, Canghong Shi, Xiaojie Li

We firstly represent only face regions using the latent variable as the domain knowledge and combine it with the non-face parts textures to generate high-quality face images with plausible contents.

Facial Inpainting

Detector-in-Detector: Multi-Level Analysis for Human-Parts

2 code implementations19 Feb 2019 Xiaojie Li, Lu Yang, Qing Song, Fuqiang Zhou

In particular, we adopt a region-based object detection structure with two carefully designed detectors to separately pay attention to the human body and body parts in a coarse-to-fine manner, which we call Detector-in-Detector network (DID-Net).

Body Detection Face Detection +2

Cannot find the paper you are looking for? You can Submit a new open access paper.