no code implementations • 5 Feb 2025 • Zhihong Xu, Dongxia Wang, Peng Du, Yang Cao, Qing Guo
To address this limitation, we propose human-parsing-guided attention diffusion, a novel approach that effectively preserves both facial and clothing appearance while generating high-quality results.
1 code implementation • 24 Jan 2025 • Runyi Hu, Jie Zhang, Yiming Li, Jiwei Li, Qing Guo, Han Qiu, Tianwei Zhang
Artificial Intelligence Generated Content (AIGC) has advanced significantly, particularly with the development of video generation models such as text-to-video (T2V) models and image-to-video (I2V) models.
no code implementations • 21 Dec 2024 • JiaYi Zhu, Qing Guo, Felix Juefei-Xu, Yihao Huang, Yang Liu, Geguang Pu
The task of co-saliency object detection (Co-SOD) seeks to identify common, salient objects across a collection of images by examining shared visual features.
no code implementations • 17 Dec 2024 • Qi Zhou, Tianlin Li, Qing Guo, Dongxia Wang, Yun Lin, Yang Liu, Jin Song Dong
Instead of directly using responses from partial images for voting, we investigate using them to supervise the LVLM's responses to the original images.
no code implementations • 13 Dec 2024 • Runyi Hu, Jie Zhang, Yiming Li, Jiwei Li, Qing Guo, Han Qiu, Tianwei Zhang
For extraction, the process is reversed: the watermarked image is inverted back to the initial watermarked noise via DDIM Inversion, from which the embedded watermark is extracted.
no code implementations • 11 Dec 2024 • Zhongyi Zhang, Jie Zhang, Wenbo Zhou, Xinghui Zhou, Qing Guo, Weiming Zhang, Tianwei Zhang, Nenghai Yu
Face-swapping techniques have advanced rapidly with the evolution of deep learning, leading to widespread use and growing concerns about potential misuse, especially in cases of fraud.
no code implementations • 11 Dec 2024 • Yun Xing, Nhat Chung, Jie Zhang, Yue Cao, Ivor Tsang, Yang Liu, Lei Ma, Qing Guo
We validate our method on both digital and physical level, \ie, nuImage and manually captured real scenes, where both statistical and visual results prove that our MAGIC is powerful and effectively for attacking wide-used object detection systems.
no code implementations • 28 Nov 2024 • Yue Cao, Yun Xing, Jie Zhang, Di Lin, Tianwei Zhang, Ivor Tsang, Yang Liu, Qing Guo
In this paper, we present the first approach to generate scene-coherent typographic adversarial attacks that mislead advanced LVLMs while maintaining visual naturalness through the capability of the LLM-based agent.
no code implementations • 11 Nov 2024 • Xingrui Yu, Zhenglin Wan, David Mark Bossens, Yueming Lyu, Qing Guo, Ivor W. Tsang
Learning diverse and high-performance behaviors from a limited set of demonstrations is a grand challenge.
2 code implementations • 4 Nov 2024 • Xiaojun Jia, Sensen Gao, Qing Guo, Ke Ma, Yihao Huang, Simeng Qin, Yang Liu, Ivor Tsang Fellow, Xiaochun Cao
Hence, we propose to generate AEs in the semantic image-text feature contrast space, which can project the original feature space into a semantic corpus subspace.
no code implementations • 8 Oct 2024 • Zhenglin Wan, Xingrui Yu, David Mark Bossens, Yueming Lyu, Qing Guo, Flint Xiaofeng Fan, Ivor Tsang
Imitation learning (IL) has shown great potential in various applications, such as robot control.
no code implementations • 25 Aug 2024 • Sensen Gao, Xiaojun Jia, Yihao Huang, Ranjie Duan, Jindong Gu, Yang Bai, Yang Liu, Qing Guo
Text-to-Image(T2I) models have achieved remarkable success in image generation and editing, yet these models still have many potential issues, particularly in generating inappropriate or Not-Safe-For-Work(NSFW) content.
no code implementations • 6 Aug 2024 • Aishan Liu, Yuguang Zhou, Xianglong Liu, Tianyuan Zhang, Siyuan Liang, Jiakai Wang, Yanjun Pu, Tianlin Li, Junqi Zhang, Wenbo Zhou, Qing Guo, DaCheng Tao
To enable context-dependent behaviors in downstream agents, we implement a dual-modality activation strategy that controls both the generation and execution of program defects through textual and visual triggers.
no code implementations • 24 Jul 2024 • Siyu Chen, Qing Guo
Employing a comprehensive survey of micro and small enterprises (MSEs) and the Digital Financial Inclusion Index in China, this study investigates the influence of fintech on MSE innovation empirically.
no code implementations • 23 Jul 2024 • Liang Zhao, Qing Guo, Xiaoguang Li, Song Wang
In this work, we identify the visual-text inpainting task to achieve high-quality scene text image restoration and text completion: Given a scene text image with unknown missing regions and the corresponding text with unknown missing characters, we aim to complete the missing information in both images and text by leveraging their complementary information.
1 code implementation • 9 Jul 2024 • Ruofei Wang, Qing Guo, Haoliang Li, Renjie Wan
However, research into the potential risk associated with backdoor attacks in asynchronous event data has been scarce, leaving related tasks vulnerable to potential threats.
Ranked #1 on
Event data classification
on N-Caltech 101
no code implementations • 22 Jun 2024 • Tianyu Wei, Shanmin Pang, Qi Guo, Yizhuo Ma, Qing Guo
In this work, we unveil a previously unrecognized and latent risk of using diffusion models to generate images; we utilize emotion in the input texts to introduce negative contents, potentially eliciting unfavorable emotions in users.
no code implementations • 10 Jun 2024 • Yihao Huang, Qing Guo, Felix Juefei-Xu, Ming Hu, Xiaojun Jia, Xiaochun Cao, Geguang Pu, Yang Liu
Universal adversarial perturbation (UAP), also known as image-agnostic perturbation, is a fixed perturbation map that can fool the classifier with high probabilities on arbitrary images, making it more practical for attacking deep models in the real world.
no code implementations • 28 May 2024 • Di Yang, Yihao Huang, Qing Guo, Felix Juefei-Xu, Xiaojun Jia, Run Wang, Geguang Pu, Yang Liu
The widespread use of diffusion methods enables the creation of highly realistic images on demand, thereby posing significant risks to the integrity and safety of online information and highlighting the necessity of DeepFake detection.
1 code implementation • 28 May 2024 • BoWen Zhang, Xiaofei Xie, Haotian Lu, Na Ma, Tianlin Li, Qing Guo
The core challenge lies in generating smooth and natural transitions between these segments given the inherent complexity and variability of action transitions.
1 code implementation • 27 May 2024 • YuXiao Lee, Xiaofeng Cao, Jingcai Guo, Wei Ye, Qing Guo, Yi Chang
The remarkable achievements of Large Language Models (LLMs) have captivated the attention of both academia and industry, transcending their initial role in dialogue generation.
no code implementations • 24 May 2024 • Qing Guo, Siyu Chen, Xiangquan Zeng
The proliferation of internet technology has catalyzed the rapid development of digital finance, significantly impacting the optimization of resource allocation in China and exerting a substantial and enduring influence on the structure of employment and income distribution.
no code implementations • 23 May 2024 • Yihao Huang, Chong Wang, Xiaojun Jia, Qing Guo, Felix Juefei-Xu, Jian Zhang, Geguang Pu, Yang Liu
With the rising popularity of Large Language Models (LLMs), assessing their trustworthiness through security tasks has gained critical importance.
no code implementations • 23 May 2024 • Nhat Chung, Sensen Gao, Tuan-Anh Vu, Jie Zhang, Aishan Liu, Yun Lin, Jin Song Dong, Qing Guo
To further explore the risk in AD systems and the transferability of practical threats, we propose to leverage typographic attacks against AD systems relying on the decision-making capabilities of Vision-LLMs.
1 code implementation • 16 Apr 2024 • Qi Guo, Shanmin Pang, Xiaojun Jia, Yang Liu, Qing Guo
Furthermore, compared to previous transfer-based adversarial attacks, the adversarial examples generated by our method have better transferability.
1 code implementation • 9 Apr 2024 • Jianlang Chen, Xuhong Ren, Qing Guo, Felix Juefei-Xu, Di Lin, Wei Feng, Lei Ma, Jianjun Zhao
To achieve high accuracy on both clean and adversarial data, we propose building a spatial-temporal continuous representation using the semantic text guidance of the object of interest.
no code implementations • CVPR 2024 • Jinlong Li, Baolu Li, Zhengzhong Tu, Xinyu Liu, Qing Guo, Felix Juefei-Xu, Runsheng Xu, Hongkai Yu
Vision-centric perception systems for autonomous driving have gained considerable attention recently due to their cost-effectiveness and scalability, especially compared to LiDAR-based systems.
no code implementations • CVPR 2024 • JiaYi Zhu, Qing Guo, Felix Juefei-Xu, Yihao Huang, Yang Liu, Geguang Pu
In this paper, we propose a novel robustness enhancement framework by first learning the concept of the co-salient objects based on the input group images and then leveraging this concept to purify adversarial perturbations, which are subsequently fed to CoSODs for robustness enhancement.
1 code implementation • 19 Mar 2024 • Sensen Gao, Xiaojun Jia, Xuhong Ren, Ivor Tsang, Qing Guo
Vision-language pre-training (VLP) models exhibit remarkable capabilities in comprehending both images and text, yet they remain susceptible to multimodal adversarial examples (AEs).
no code implementations • 19 Feb 2024 • Tianlin Li, Qian Liu, Tianyu Pang, Chao Du, Qing Guo, Yang Liu, Min Lin
The emerging success of large language models (LLMs) heavily relies on collecting abundant training data from external (untrusted) sources.
no code implementations • 19 Feb 2024 • Tianlin Li, XiaoYu Zhang, Chao Du, Tianyu Pang, Qian Liu, Qing Guo, Chao Shen, Yang Liu
Building on this insight and observation, we develop FairThinking, a pipeline designed to automatically generate roles that enable LLMs to articulate diverse perspectives for fair expressions.
no code implementations • 6 Feb 2024 • Bohao Qu, Xiaofeng Cao, Qing Guo, Yi Chang, Ivor W. Tsang, Chengqi Zhang
In this study, we present a transductive inference approach on that reward information propagation graph, which enables the effective estimation of rewards for unlabelled data in offline reinforcement learning.
no code implementations • 6 Feb 2024 • Qi Zhou, Dongxia Wang, Tianlin Li, Zhihong Xu, Yang Liu, Kui Ren, Wenhai Wang, Qing Guo
To expose this potential vulnerability, we aim to build an adversarial attack forcing SDEdit to generate a specific data distribution aligned with a specified attribute (e. g., female), without changing the input's attribute characteristics.
no code implementations • 5 Feb 2024 • Yihao Huang, Kaiyuan Yu, Qing Guo, Felix Juefei-Xu, Xiaojun Jia, Tianlin Li, Geguang Pu, Yang Liu
In recent years, LiDAR-camera fusion models have markedly advanced 3D object detection tasks in autonomous driving.
1 code implementation • 30 Jan 2024 • Jinlong Li, Baolu Li, Xinyu Liu, Jianwu Fang, Felix Juefei-Xu, Qing Guo, Hongkai Yu
The multi-agent perception system collects visual data from sensors located on various agents and leverages their relative poses determined by GPS signals to effectively fuse information, mitigating the limitations of single-agent sensing, such as occlusion.
1 code implementation • 4 Jan 2024 • Ruofei Wang, Renjie Wan, Zongyu Guo, Qing Guo, Rui Huang
Backdoor attack aims to deceive a victim model when facing backdoor instances while maintaining its performance on benign data.
no code implementations • 29 Dec 2023 • Tuan-Anh Vu, Duc Thanh Nguyen, Qing Guo, Binh-Son Hua, Nhat Minh Chung, Ivor W. Tsang, Sai-Kit Yeung
Such cross-domain representations are desirable in segmenting camouflaged objects where visual cues are subtle to distinguish the objects from the background, especially in segmenting novel objects which are not seen in training.
1 code implementation • 5 Dec 2023 • Rui Huang, Binbin Jiang, Qingyi Zhao, William Wang, Yuxiang Zhang, Qing Guo
Our approach surpasses state-of-the-art 2D change detection and NeRF-based methods by a significant margin.
no code implementations • 3 Dec 2023 • Xiaojun Jia, Jindong Gu, Yihao Huang, Simeng Qin, Qing Guo, Yang Liu, Xiaochun Cao
At the second stage, the pixels are divided into different branches based on their transferable property which is dependent on Kullback-Leibler divergence.
1 code implementation • 18 Oct 2023 • Yue Cao, Tianlin Li, Xiaofeng Cao, Ivor Tsang, Yang Liu, Qing Guo
The underlying rationale behind our idea is that image resampling can alleviate the influence of adversarial perturbations while preserving essential semantic information, thereby conferring an inherent advantage in defending against adversarial attacks.
no code implementations • 13 Oct 2023 • Canyu Zhang, Xiaoguang Li, Qing Guo, Song Wang
To this end, we propose a framework with two modules: (1) building a semantic implicit representation (SIR) for a corrupted image whose large regions miss.
no code implementations • 3 Aug 2023 • Kairui Yang, Enhui Ma, Jibin Peng, Qing Guo, Di Lin, Kaicheng Yu
To this end, we propose a two-stage generative method, dubbed BEVControl, that can generate accurate foreground and background contents.
1 code implementation • ICCV 2023 • Rabab Abdelfattah, Qing Guo, Xiaoguang Li, XiaoFeng Wang, Song Wang
Using the aggregated similarity scores as the initial pseudo labels at the training stage, we propose an optimization framework to train the parameters of the classification network and refine pseudo labels for unobserved labels.
no code implementations • 26 Jul 2023 • Canyu Zhang, Qing Guo, Xiaoguang Li, Renjie Wan, Hongkai Yu, Ivor Tsang, Song Wang
Given the coordinates of a pixel we want to reconstruct, we first collect its neighboring pixels in the input image and extract their detail-enhanced semantic embeddings, unmask-attentional semantic embeddings, importance values, and spatial distances to the desired pixel.
1 code implementation • ICCV 2023 • Ziyuan Luo, Qing Guo, Ka Chun Cheung, Simon See, Renjie Wan
Neural Radiance Fields (NeRF) have the potential to be a major representation of media.
no code implementations • ICCV 2023 • Haotian Dong, Enhui Ma, Lubo Wang, Miaohui Wang, Wuyuan Xie, Qing Guo, Ping Li, Lingyu Liang, Kairui Yang, Di Lin
In this paper, we propose Cross-View Synthesis Transformer (CVSformer), which consists of Multi-View Feature Synthesis and Cross-View Transformer for learning cross-view object relationships.
no code implementations • 30 Jun 2023 • Huiming Sun, Lan Fu, Jinlong Li, Qing Guo, Zibo Meng, Tianyun Zhang, Yuewei Lin, Hongkai Yu
Furthermore, we design DefenseNet as a learn-able pre-processing to the adversarial cloudy images so as to preserve the performance of the deep learning based remote sensing SOD model, without tuning the already deployed deep SOD model.
no code implementations • 27 Jun 2023 • Tianlin Li, Qing Guo, Aishan Liu, Mengnan Du, Zhiming Li, Yang Liu
Existing fairness regularization terms fail to achieve decision rationale alignment because they only constrain last-layer outputs while ignoring intermediate neuron alignment.
no code implementations • 22 Jun 2023 • Jin Ma, Jinlong Li, Qing Guo, Tianyun Zhang, Yuewei Lin, Hongkai Yu
The emergence of different sensors (Near-Infrared, Depth, etc.)
no code implementations • 25 May 2023 • Yihao Huang, Yue Cao, Tianlin Li, Felix Juefei-Xu, Di Lin, Ivor W. Tsang, Yang Liu, Qing Guo
Second, we extend representative adversarial attacks against SAM and study the influence of different prompts on robustness.
no code implementations • 18 May 2023 • Xiaoguang Li, Qing Guo, Pingping Cai, Wei Feng, Ivor Tsang, Song Wang
State-of-the-art shadow removal methods train deep neural networks on collected shadow & shadow-free image pairs, which are desired to complete two distinct tasks via shared weights, i. e., data restoration for shadow regions and identical mapping for non-shadow regions.
no code implementations • 18 May 2023 • Di Yang, Yihao Huang, Qing Guo, Felix Juefei-Xu, Ming Hu, Yang Liu, Geguang Pu
The adversarial patch attack aims to fool image classifiers within a bounded, contiguous region of arbitrary changes, posing a real threat to computer vision systems (e. g., autonomous driving, content moderation, biometric authentication, medical imaging) in the physical world.
no code implementations • 18 May 2023 • Yihao Huang, Felix Juefei-Xu, Qing Guo, Jie Zhang, Yutong Wu, Ming Hu, Tianlin Li, Geguang Pu, Yang Liu
Although recent personalization methods have democratized high-resolution image synthesis by enabling swift concept acquisition with minimal examples and lightweight computation, they also present an exploitable avenue for high accessible backdoor attacks.
no code implementations • CVPR 2023 • Yang Hou, Qing Guo, Yihao Huang, Xiaofei Xie, Lei Ma, Jianjun Zhao
Second, we find that the statistical differences between natural and DeepFake images are positively associated with the distribution shifting between the two kinds of images, and we propose to use a distribution-aware loss to guide the optimization of different degradations.
no code implementations • 21 Feb 2023 • Anran Li, Hongyi Peng, Lan Zhang, Jiahui Huang, Qing Guo, Han Yu, Yang Liu
Vertical Federated Learning (VFL) enables multiple data owners, each holding a different subset of features about largely overlapping sets of data sample(s), to jointly train a useful global model.
1 code implementation • ICCV 2023 • Xiaoguang Li, Qing Guo, Rabab Abdelfattah, Di Lin, Wei Feng, Ivor Tsang, Song Wang
In this work, we find that pretraining shadow removal networks on the image inpainting dataset can reduce the shadow remnants significantly: a naive encoder-decoder network gets competitive restoration quality w. r. t.
no code implementations • 9 Jan 2023 • Yuhao Liu, Qing Guo, Lan Fu, Zhanghan Ke, Ke Xu, Wei Feng, Ivor W. Tsang, Rynson W. H. Lau
Hence, in this paper, we propose to remove shadows at the image structure level.
no code implementations • CVPR 2023 • Leming Guo, Wanli Xue, Qing Guo, Bo Liu, Kaihua Zhang, Tiantian Yuan, ShengYong Chen
Existing results in [9, 20, 25, 36] have indicated that, as the frontal component of the overall model, the spatial perception module used for spatial feature extraction tends to be insufficiently trained.
Ranked #7 on
Sign Language Recognition
on CSL-Daily
no code implementations • 28 Nov 2022 • Linkun Fan, Fazhi He, Qing Guo, Wei Tang, Xiaolin Hong, Bing Li
As a result, backdoor pattern designed for one certain 3D data structure will be disable for other data structures of the same 3D scene.
1 code implementation • 21 Nov 2022 • Rui Huang, Ruofei Wang, Qing Guo, Jieda Wei, Yuxiang Zhang, Wei Fan, Yang Liu
Change detection (CD) is to decouple object changes (i. e., object missing or appearing) from background changes (i. e., environment variations) like light and season variations in two images captured in the same scene over a long time span, presenting critical applications in disaster management, urban development, etc.
no code implementations • 20 Nov 2022 • Huiming Sun, Jin Ma, Qing Guo, Qin Zou, Shaoyue Song, Yuewei Lin, Hongkai Yu
To the best of our knowledge, all the existing image inpainting algorithms learn to repair the occluded regions for a better visualization quality, they are excellent for natural images but not good enough for geoscience images by ignoring the geoscience related tasks.
no code implementations • 12 Oct 2022 • Shuangzhi Li, Zhijie Wang, Felix Juefei-Xu, Qing Guo, Xingyu Li, Lei Ma
Then, for the first attempt, we construct a benchmark based on the physical-aware common corruptions for point cloud detectors, which contains a total of 1, 122, 150 examples covering 7, 481 scenes, 25 common corruption types, and 6 severities.
no code implementations • 21 Sep 2022 • Xuhong Ren, Jianlang Chen, Felix Juefei-Xu, Wanli Xue, Qing Guo, Lei Ma, Jianjun Zhao, ShengYong Chen
Then, we propose a novel core-failure-set guided DARTS that embeds a K-center-greedy algorithm for DARTS to select suitable corrupted failure examples to refine the model architecture.
1 code implementation • 15 Jul 2022 • Qing Guo, Ruofei Wang, Rui Huang, Shuifa Sun, Yuxiang Zhang
Change detection (CD) aims to detect change regions within an image pair captured at different times, playing a significant role in diverse real-world applications.
no code implementations • 25 Apr 2022 • Ruize Han, Wei Feng, Qing Guo, QinGhua Hu
Visual object tracking is an important task in computer vision, which has many real-world applications, e. g., video surveillance, visual navigation.
no code implementations • 24 Mar 2022 • Xiaofei Xie, Tianlin Li, Jian Wang, Lei Ma, Qing Guo, Felix Juefei-Xu, Yang Liu
Inspired by software testing, a number of structural coverage criteria are designed and proposed to measure the test adequacy of DNNs.
1 code implementation • CVPR 2022 • Xiaoguang Li, Qing Guo, Di Lin, Ping Li, Wei Feng, Song Wang
As a result, the final method takes the advantage of effective semantic & image-level filling for high-fidelity inpainting.
no code implementations • 17 Jan 2022 • JiaYi Zhu, Qing Guo, Felix Juefei-Xu, Yihao Huang, Yang Liu, Geguang Pu
Modern face recognition systems (FRS) still fall short when the subjects are wearing facial masks, a common theme in the age of respiratory pandemics.
no code implementations • 16 Jan 2022 • Yihao Huang, Liangru Sun, Qing Guo, Felix Juefei-Xu, JiaYi Zhu, Jincao Feng, Yang Liu, Geguang Pu
To obtain adversarial examples with a high attack success rate, we propose unconstrained enhancement in terms of the light and shade relationship in images.
1 code implementation • 7 Jan 2022 • Qing Guo, Jingyang Sun, Felix Juefei-Xu, Lei Ma, Di Lin, Wei Feng, Song Wang
First, we propose the uncertainty-aware cascaded predictive filtering (UC-PFilt) that can identify the difficulties of reconstructing clean pixels via predicted kernels and remove the residual rain traces effectively.
no code implementations • 27 Nov 2021 • Lan Fu, Qing Guo, Felix Juefei-Xu, Hongkai Yu, Wei Feng, Yang Liu, Song Wang
The observation of this work motivates us to design a novel detection-aware shadow removal framework, which empowers shadow removal to achieve higher restoration quality and enhance the shadow robustness of deployed facial landmark detectors.
no code implementations • 26 Nov 2021 • Hua Qi, Zhijie Wang, Qing Guo, Jianlang Chen, Felix Juefei-Xu, Lei Ma, Jianjun Zhao
In this work, as the first attempt, we initiate to repair DNNs by jointly optimizing the architecture and weights at a higher (i. e., block) level.
no code implementations • 25 Nov 2021 • Yihao Huang, Felix Juefei-Xu, Qing Guo, Geguang Pu, Yang Liu
Bokeh effect is a natural shallow depth-of-field phenomenon that blurs the out-of-focus part in photography.
no code implementations • 18 Aug 2021 • Qian Zhang, Qing Guo, Ruijun Gao, Felix Juefei-Xu, Hongkai Yu, Wei Feng
To this end, we first propose the physical modelbased adversarial relighting attack (ARA) denoted as albedoquotient-based adversarial relighting attack (AQ-ARA).
no code implementations • 28 Jul 2021 • Qing Guo, Zhijie Wang, Felix Juefei-Xu, Di Lin, Lei Ma, Wei Feng, Yang Liu
3D point cloud completion is very challenging because it heavily relies on the accurate understanding of the complex 3D shapes (e. g., high-curvature, concave/convex, and hollowed-out 3D shapes) and the unknown & diverse patterns of the partially available point clouds.
1 code implementation • ICCV 2021 • Qing Guo, Ziyi Cheng, Felix Juefei-Xu, Lei Ma, Xiaofei Xie, Yang Liu, Jianjun Zhao
In this work, we explore the robustness of visual object trackers against motion blur from a new angle, i. e., adversarial blur attack (ABA).
no code implementations • 14 Jul 2021 • Yihao Huang, Qing Guo, Felix Juefei-Xu, Lei Ma, Weikai Miao, Yang Liu, Geguang Pu
To this end, we first comprehensively investigate two kinds of pixel denoising methods for adversarial robustness enhancement (i. e., existing additive-based and unexplored filtering-based methods) under the loss functions of image-level and semantic-level, respectively, showing that pixel-wise filtering can obtain much higher image quality (e. g., higher PSNR) as well as higher robustness (e. g., higher accuracy on adversarial examples) than existing pixel-wise additive-based method.
1 code implementation • 9 Jul 2021 • Qing Guo, Xiaoguang Li, Felix Juefei-Xu, Hongkai Yu, Yang Liu, Song Wang
In this paper, for the first time, we formulate image inpainting as a mix of two problems, predictive filtering and deep generation.
1 code implementation • 2 Jul 2021 • Qing Guo, Junya Chen, Dong Wang, Yuewei Yang, Xinwei Deng, Lawrence Carin, Fan Li, Jing Huang, Chenyang Tao
Successful applications of InfoNCE and its variants have popularized the use of contrastive variational mutual information (MI) estimators in machine learning.
1 code implementation • 2 Jul 2021 • Junya Chen, Zhe Gan, Xuan Li, Qing Guo, Liqun Chen, Shuyang Gao, Tagyoung Chung, Yi Xu, Belinda Zeng, Wenlian Lu, Fan Li, Lawrence Carin, Chenyang Tao
InfoNCE-based contrastive representation learners, such as SimCLR, have been tremendously successful in recent years.
no code implementations • 18 May 2021 • Qing Guo, Felix Juefei-Xu, Changqing Zhou, Wei Feng, Yang Liu, Song Wang
In both cases, Sparta leads to CNNs with higher robustness than the vanilla ReLU, verifying the flexibility and versatility of the proposed method.
1 code implementation • 12 May 2021 • Binyu Tian, Felix Juefei-Xu, Qing Guo, Xiaofei Xie, Xiaohong Li, Yang Liu
Moreover, we propose the geometry-aware level-set optimization method to solve the adversarial vignetting regions and physical parameters jointly.
no code implementations • 11 May 2021 • Lan Fu, Hongkai Yu, Felix Juefei-Xu, Jinlong Li, Qing Guo, Song Wang
As one of the state-of-the-art perception approaches, detecting the interested objects in each frame of video surveillance is widely desired by ITS.
no code implementations • 28 Apr 2021 • Ruijun Gao, Qing Guo, Felix Juefei-Xu, Hongkai Yu, Wei Feng
We also visualize the correlation matrices, which inspire us to jointly apply different perturbations to improve the success rate of the attack.
no code implementations • 23 Apr 2021 • Ziyi Cheng, Xuhong Ren, Felix Juefei-Xu, Wanli Xue, Qing Guo, Lei Ma, Jianjun Zhao
Online updating of the object model via samples from historical frames is of great importance for accurate visual object tracking.
2 code implementations • CVPR 2021 • Lan Fu, Changqing Zhou, Qing Guo, Felix Juefei-Xu, Hongkai Yu, Wei Feng, Yang Liu, Song Wang
We conduct extensive experiments on the ISTD, ISTD+, and SRD datasets to validate our method's effectiveness and show better performance in shadow regions and comparable performance in non-shadow regions over the state-of-the-art methods.
Ranked #6 on
Shadow Removal
on ISTD+
1 code implementation • 27 Feb 2021 • Felix Juefei-Xu, Run Wang, Yihao Huang, Qing Guo, Lei Ma, Yang Liu
To fill this gap, in this paper, we provide a comprehensive overview and detailed analysis of the research work on the topic of DeepFake generation, DeepFake detection as well as evasion of DeepFake detection, with more than 318 research papers carefully surveyed.
no code implementations • 1 Jan 2021 • Qing Guo, Felix Juefei-Xu, Changqing Zhou, Lei Ma, Xiaofei Xie, Wei Feng, Yang Liu
Moreover, comprehensive evaluations have demonstrated two important properties of our method: First, superior transferability across DNNs.
no code implementations • 19 Nov 2020 • Bing Yu, Hua Qi, Qing Guo, Felix Juefei-Xu, Xiaofei Xie, Lei Ma, Jianjun Zhao
In this paper, we propose a style-guided data augmentation for repairing DNN in the operational environment.
no code implementations • 19 Sep 2020 • Binyu Tian, Qing Guo, Felix Juefei-Xu, Wen Le Chan, Yupeng Cheng, Xiaohong Li, Xiaofei Xie, Shengchao Qin
Our method reveals the potential threat to the DNN-based X-ray automated diagnosis and can definitely benefit the development of bias-field-robust automated diagnosis system.
no code implementations • 19 Sep 2020 • Yihao Huang, Felix Juefei-Xu, Qing Guo, Yang Liu, Geguang Pu
We first demonstrate that frequency-domain notch filtering, although famously shown to be effective in removing periodic noise in the spatial domain, is infeasible for our task at hand due to the manual designs required for the notch filters.
no code implementations • 19 Sep 2020 • Liming Zhai, Felix Juefei-Xu, Qing Guo, Xiaofei Xie, Lei Ma, Wei Feng, Shengchao Qin, Yang Liu
To defend the DNNs from the negative rain effect, we also present a defensive deraining strategy, for which we design an adversarial rain augmentation that uses mixed adversarial rain layers to enhance deraining models for downstream DNN perception.
1 code implementation • CVPR 2022 • Ruijun Gao, Qing Guo, Felix Juefei-Xu, Hongkai Yu, Huazhu Fu, Wei Feng, Yang Liu, Song Wang
In this paper, we address this problem from the perspective of adversarial attacks and identify a novel task: adversarial co-saliency attack.
2 code implementations • 19 Sep 2020 • Qing Guo, Jingyang Sun, Felix Juefei-Xu, Lei Ma, Xiaofei Xie, Wei Feng, Yang Liu
To fill this gap, in this paper, we regard the single-image deraining as a general image-enhancing problem and originally propose a model-free deraining method, i. e., EfficientDeRain, which is able to process a rainy image within 10~ms (i. e., around 6~ms on average), over 80 times faster than the state-of-the-art method (i. e., RCDNet), while achieving similar de-rain effects.
no code implementations • 19 Sep 2020 • Yupeng Cheng, Qing Guo, Felix Juefei-Xu, Huazhu Fu, Shang-Wei Lin, Weisi Lin
Diabetic Retinopathy (DR) is a leading cause of vision loss around the world.
no code implementations • 14 Jul 2020 • Yupeng Cheng, Qing Guo, Felix Juefei-Xu, Wei Feng, Shang-Wei Lin, Weisi Lin, Yang Liu
To this end, we initiate the very first attempt to study this problem from the perspective of adversarial attack and propose the adversarial denoise attack.
1 code implementation • 13 Jun 2020 • Yihao Huang, Felix Juefei-Xu, Run Wang, Qing Guo, Lei Ma, Xiaofei Xie, Jianwen Li, Weikai Miao, Yang Liu, Geguang Pu
At this moment, GAN-based image generation methods are still imperfect, whose upsampling design has limitations in leaving some certain artifact patterns in the synthesized image.
no code implementations • 13 Jun 2020 • Hua Qi, Qing Guo, Felix Juefei-Xu, Xiaofei Xie, Lei Ma, Wei Feng, Yang Liu, Jianjun Zhao
As the GAN-based face image and video generation techniques, widely known as DeepFakes, have become more and more matured and realistic, there comes a pressing and urgent demand for effective DeepFakes detectors.
1 code implementation • NeurIPS 2020 • Qing Guo, Felix Juefei-Xu, Xiaofei Xie, Lei Ma, Jian Wang, Bing Yu, Wei Feng, Yang Liu
Besides, the attack is further enhanced by adaptively tuning the translations of object and background.
no code implementations • 27 Jan 2020 • Yihao Huang, Felix Juefei-Xu, Qing Guo, Yang Liu, Geguang Pu
In this work, we investigate the architecture of existing GAN-based face manipulation methods and observe that the imperfection of upsampling methods therewithin could be served as an important asset for GAN-synthesized fake image detection and forgery localization.
no code implementations • 9 Dec 2019 • Run Wang, Felix Juefei-Xu, Qing Guo, Yihao Huang, Xiaofei Xie, Lei Ma, Yang Liu
In this paper, we investigate and introduce a new type of adversarial attack to evade FR systems by manipulating facial content, called \textbf{\underline{a}dversarial \underline{mor}phing \underline{a}ttack} (a. k. a.
1 code implementation • ECCV 2020 • Qing Guo, Xiaofei Xie, Felix Juefei-Xu, Lei Ma, Zhongguo Li, Wanli Xue, Wei Feng, Yang Liu
We identify that online object tracking poses two new challenges: 1) it is difficult to generate imperceptible perturbations that can transfer across frames, and 2) real-time trackers require the attack to satisfy a certain level of efficiency.
no code implementations • 19 Sep 2019 • Zhu Sun, Qing Guo, Jie Yang, Hui Fang, Guibing Guo, Jie Zhang, Robin Burke
This Research Commentary aims to provide a comprehensive and systematic survey of the recent research on recommender systems with side information.
no code implementations • 21 Aug 2019 • Qing Guo, Wei Feng, Zhihao Chen, Ruijun Gao, Liang Wan, Song Wang
In this paper, we address these two problems by constructing a Blurred Video Tracking benchmark, which contains a variety of videos with different levels of motion blurs, as well as ground truth tracking results for evaluating trackers.
no code implementations • ICCV 2017 • Qing Guo, Wei Feng, Ce Zhou, Rui Huang, Liang Wan, Song Wang
How to effectively learn temporal variation of target appearance, to exclude the interference of cluttered background, while maintaining real-time response, is an essential problem of visual object tracking.
Ranked #5 on
Visual Object Tracking
on OTB-2013