no code implementations • 2 Dec 2024 • Delong Zhu, Yuezun Li, Baoyuan Wu, Jiaran Zhou, Zhibo Wang, Siwei Lyu
The motivation stems from the reliance of most DeepFake methods on face detectors to automatically extract victim faces from videos for training or synthesis (testing).
no code implementations • 29 Nov 2024 • Xinjie Cui, Yuezun Li, Ao Luo, Jiaran Zhou, Junyu Dong
We describe the Forensics Adapter, an adapter network designed to transform CLIP into an effective and generalizable face forgery detector.
no code implementations • 29 Oct 2024 • Xudong Wang, Yuezun Li, Huiyu Zhou, Jiaran Zhou, Junyu Dong
Subsequently, we describe a structural-agnostic graph reasoning strategy tailored for our graph to enhance the representation of nodes.
no code implementations • 22 Oct 2024 • Yang Yu, Yuezun Li, Xin Sun, Junyu Dong
Phytoplankton are a crucial component of aquatic ecosystems, and effective monitoring of them can provide valuable insights into ocean environments and ecosystem changes.
no code implementations • 10 Oct 2024 • Dongliang Zhang, Yunfei Li, Jiaran Zhou, Yuezun Li
These varying qualities diversify the pattern of forgery traces, significantly increasing the difficulty of DeepFake detection.
no code implementations • 22 Sep 2024 • Yuzhen Lin, Wentang Song, Bin Li, Yuezun Li, Jiangqun Ni, Han Chen, Qiushi Li
Previous studies in deepfake detection have shown promising results when testing face forgeries from the same dataset as the training.
no code implementations • 5 Sep 2024 • Pu Sun, Honggang Qi, Yuezun Li
However, these methods mainly concentrate on capturing the blending inconsistency in DeepFake faces, raising a new security issue, termed Active Fake, emerges when individuals intentionally create blending inconsistency in their authentic videos to evade responsibility.
no code implementations • 3 Sep 2024 • Qingxuan Lv, Junyu Dong, Yuezun Li, Sheng Chen, Hui Yu, Shu Zhang, Wenhan Wang
To enable further advance in underwater stereo matching, we introduce a large synthetic dataset called UWStereo.
no code implementations • 29 Aug 2024 • Yangxiang Zhang, Yuezun Li, Ao Luo, Jiaran Zhou, Junyu Dong
In this paper, we describe an efficient two-stream architecture for real-time image manipulation detection.
no code implementations • 29 Jun 2024 • Yang Yu, Qingxuan Lv, Yuezun Li, Zhiqiang Wei, Junyu Dong
Phytoplankton, a crucial component of aquatic ecosystems, requires efficient monitoring to understand marine ecological processes and environmental conditions.
no code implementations • 11 May 2024 • Jinkun Jiang, Qingxuan Lv, Yuezun Li, Yong Du, Sheng Chen, Hui Yu, Junyu Dong
The drawback of these methods includes: 1) the pair-wise relation is limited to exposing the underlying correlations of two more samples, hindering the exploration of the structural information embedded in the target domain; 2) the clustering process only relies on the semantic feature, while overlooking the critical effect of domain shift, i. e., the distribution differences between the source and target domains.
no code implementations • 22 Apr 2024 • Yunfei Li, Yuezun Li, Xin Wang, Baoyuan Wu, Jiaran Zhou, Junyu Dong
Our method features four major improvements: \ding{182} we describe a new texture-aware branch that effectively captures subtle manipulation traces with a Diversiform Pixel Difference Attention module.
no code implementations • 22 Apr 2024 • Hanzhe Li, Jiaran Zhou, Yuezun Li, Baoyuan Wu, Bin Li, Junyu Dong
Existing methods typically generate these faces by blending real or fake faces in spatial domain.
no code implementations • 17 Apr 2024 • Ying Zhang, Yuezun Li, Bo Peng, Jiaran Zhou, Huiyu Zhou, Junyu Dong
The task of video inpainting detection is to expose the pixel-level inpainted regions within a video sequence.
no code implementations • 17 Dec 2023 • Qingxuan Lv, Yuezun Li, Junyu Dong, Sheng Chen, Hui Yu, Huiyu Zhou, Shu Zhang
Specifically, our strategy considers both forward and backward adaptation, to transfer the forgery knowledge from the source domain to the target domain in forward adaptation and then reverse the adaptation from the target domain to the source domain in backward adaptation.
1 code implementation • 3 Aug 2023 • Cong Zhang, Honggang Qi, Shuhui Wang, Yuezun Li, Siwei Lyu
One straightforward way to address this issue is to simultaneous process multi-face by integrating face extraction and forgery detection in an end-to-end fashion by adapting advanced object detection architectures.
no code implementations • 2 Aug 2023 • Jiucui Lu, Jiaran Zhou, Junyu Dong, Bin Li, Siwei Lyu, Yuezun Li
The proposed ForensicsForest family is composed of three variants, which are {\em ForensicsForest}, {\em Hybrid ForensicsForest} and {\em Divide-and-Conquer ForensicsForest} respectively.
no code implementations • 27 Jul 2023 • Pu Sun, Honggang Qi, Yuezun Li, Siwei Lyu
In light of these two traces, our method can effectively expose DeepFakes by identifying them.
1 code implementation • 22 Apr 2022 • Xianglong, Yuezun Li, Haipeng Qu, Junyu Dong
However, the guidance map is fixed in existing methods, which can not consistently reflect the behavior of networks as the image is changed during iteration.
no code implementations • 3 Jun 2021 • Quanyu Liao, Yuezun Li, Xin Wang, Bin Kong, Bin Zhu, Siwei Lyu, Youbing Yin, Qi Song, Xi Wu
Fooling people with highly realistic fake images generated with Deepfake or GANs brings a great social disturbance to our society.
1 code implementation • 2 Jun 2021 • Bo Peng, Hongxing Fan, Wei Wang, Jing Dong, Yuezun Li, Siwei Lyu, Qi Li, Zhenan Sun, Han Chen, Baoying Chen, Yanjie Hu, Shenghai Luo, Junrui Huang, Yutong Yao, Boyuan Liu, Hefei Ling, Guosheng Zhang, Zhiliang Xu, Changtao Miao, Changlei Lu, Shan He, Xiaoyan Wu, Wanyi Zhuang
This competition provides a common platform for benchmarking the adversarial game between current state-of-the-art DeepFake creation and detection methods.
no code implementations • 2 Mar 2021 • Yuezun Li, Cong Zhang, Pu Sun, Honggang Qi, Siwei Lyu
In recent years, the advent of deep learning-based techniques and the significant reduction in the cost of computation resulted in the feasibility of creating realistic videos of human faces, commonly known as DeepFakes.
no code implementations • 1 Feb 2021 • Pu Sun, Yuezun Li, Honggang Qi, Siwei Lyu
In this paper, we describe Landmark Breaker, the first dedicated method to disrupt facial landmark extraction, and apply it to the obstruction of the generation of DeepFake videos. Our motivation is that disrupting the facial landmark extraction can affect the alignment of input face so as to degrade the DeepFake quality.
1 code implementation • 31 Oct 2020 • Pu Sun, Yuezun Li, Honggang Qi, Siwei Lyu
Face synthesis is an important problem in computer vision with many applications.
1 code implementation • 24 Sep 2020 • Shu Hu, Yuezun Li, Siwei Lyu
We show that such artifacts exist widely in high-quality GAN synthesized faces and further describe an automatic method to extract and compare corneal specular highlights from two eyes.
no code implementations • 19 Oct 2019 • Yuezun Li, Ao Luo, Siwei Lyu
In this paper, we describe a fast and light-weight portrait segmentation method based on a new highly light-weight backbone (HLB) architecture.
8 code implementations • CVPR 2020 • Yuezun Li, Xin Yang, Pu Sun, Honggang Qi, Siwei Lyu
AI-synthesized face-swapping videos, commonly known as DeepFakes, is an emerging problem threatening the trustworthiness of online information.
no code implementations • 21 Jun 2019 • Yuezun Li, Xin Yang, Baoyuan Wu, Siwei Lyu
Recent years have seen fast development in synthesizing realistic human faces using AI technologies.
no code implementations • 30 Mar 2019 • Xin Yang, Yuezun Li, Honggang Qi, Siwei Lyu
Generative adversary networks (GANs) have recently led to highly realistic image synthesis results.
no code implementations • 12 Feb 2019 • Yuezun Li, Siwei Lyu
In this work, we describe a new face de-identification method that can preserve essential facial attributes in the faces while concealing the identities.
3 code implementations • 1 Nov 2018 • Yuezun Li, Siwei Lyu
Compared to previous methods which use a large amount of real and DeepFake generated images to train CNN classifier, our method does not need DeepFake generated images as negative training examples since we target the artifacts in affine face warping as the distinctive feature to distinguish real and fake images.
1 code implementation • 1 Nov 2018 • Xin Yang, Yuezun Li, Siwei Lyu
In this paper, we propose a new method to expose AI-generated fake face images or videos (commonly known as the Deep Fakes).
no code implementations • 16 Sep 2018 • Yuezun Li, Xiao Bian, Ming-Ching Chang, Siwei Lyu
In this paper, we focus on exploring the vulnerability of the Single Shot Module (SSM) commonly used in recent object detectors, by adding small perturbations to patches in the background outside the object.
no code implementations • 16 Sep 2018 • Yuezun Li, Daniel Tian, Ming-Ching Chang, Xiao Bian, Siwei Lyu
Adversarial noises are useful tools to probe the weakness of deep learning based computer vision algorithms.
3 code implementations • 7 Jun 2018 • Yuezun Li, Ming-Ching Chang, Siwei Lyu
The new developments in deep generative networks have significantly improve the quality and efficiency in generating realistically-looking fake face videos.