Search Results for author: Yong Guo

Found 52 papers, 32 papers with code

MambaIRv2: Attentive State Space Restoration

1 code implementation22 Nov 2024 Hang Guo, Yong Guo, Yaohua Zha, Yulun Zhang, Wenbo Li, Tao Dai, Shu-Tao Xia, Yawei Li

The Mamba-based image restoration backbones have recently demonstrated significant potential in balancing global reception and computational efficiency.

Computational Efficiency Image Restoration +1

BiDense: Binarization for Dense Prediction

1 code implementation15 Nov 2024 Rui Yin, Haotong Qin, Yulun Zhang, Wenbo Li, Yong Guo, Jianjun Zhu, Cheng Wang, Biao Jia

BiDense incorporates two key techniques: the Distribution-adaptive Binarizer (DAB) and the Channel-adaptive Full-precision Bypass (CFB).

Binarization

Hi-Mamba: Hierarchical Mamba for Efficient Image Super-Resolution

no code implementations14 Oct 2024 Junbo Qiao, Jincheng Liao, Wei Li, Yulun Zhang, Yong Guo, Yi Wen, Zhangxizi Qiu, Jiao Xie, Jie Hu, Shaohui Lin

State Space Models (SSM), such as Mamba, have shown strong representation ability in modeling long-range dependency with linear complexity, achieving successful applications from high-level to low-level vision tasks.

Image Super-Resolution Mamba +1

Gap Preserving Distillation by Building Bidirectional Mappings with A Dynamic Teacher

no code implementations5 Oct 2024 Yong Guo, Shulian Zhang, Haolin Pan, Jing Liu, Yulun Zhang, Jian Chen

To address this, we propose a Gap Preserving Distillation (GPD) method that trains an additional dynamic teacher model from scratch along with training the student to bridge this gap.

Knowledge Distillation

Distillation-Free One-Step Diffusion for Real-World Image Super-Resolution

1 code implementation5 Oct 2024 Jianze Li, JieZhang Cao, Zichen Zou, Xiongfei Su, Xin Yuan, Yulun Zhang, Yong Guo, Xiaokang Yang

However, these methods incur substantial training costs and may constrain the performance of the student model by the teacher's limitations.

Image Super-Resolution Knowledge Distillation

Effective Diffusion Transformer Architecture for Image Super-Resolution

1 code implementation29 Sep 2024 Kun Cheng, Lei Yu, Zhijun Tu, Xiao He, Liyu Chen, Yong Guo, Mingrui Zhu, Nannan Wang, Xinbo Gao, Jie Hu

In this work, we design an effective diffusion transformer for image super-resolution (DiT-SR) that achieves the visual quality of prior-based methods, but through a training-from-scratch manner.

Image Generation Image Super-Resolution

One Step Diffusion-based Super-Resolution with Time-Aware Distillation

1 code implementation14 Aug 2024 Xiao He, Huaao Tang, Zhijun Tu, Junchao Zhang, Kun Cheng, Hanting Chen, Yong Guo, Mingrui Zhu, Nannan Wang, Xinbo Gao, Jie Hu

Specifically, we introduce a novel score distillation strategy to align the data distribution between the outputs of the student and teacher models after minor noise perturbation.

Image Super-Resolution Knowledge Distillation

Enhanced Long-Tailed Recognition with Contrastive CutMix Augmentation

2 code implementations6 Jul 2024 Haolin Pan, Yong Guo, Mianjie Yu, Jian Chen

Real-world data often follows a long-tailed distribution, where a few head classes occupy most of the data and a large number of tail classes only contain very limited samples.

Contrastive Learning Data Augmentation +2

UltraPixel: Advancing Ultra-High-Resolution Image Synthesis to New Peaks

no code implementations2 Jul 2024 Jingjing Ren, Wenbo Li, Haoyu Chen, Renjing Pei, Bin Shao, Yong Guo, Long Peng, Fenglong Song, Lei Zhu

Ultra-high-resolution image generation poses great challenges, such as increased semantic planning complexity and detail synthesis difficulties, alongside substantial training resource demands.

Computational Efficiency Denoising +1

2DQuant: Low-bit Post-Training Quantization for Image Super-Resolution

1 code implementation10 Jun 2024 Kai Liu, Haotong Qin, Yong Guo, Xin Yuan, Linghe Kong, Guihai Chen, Yulun Zhang

Low-bit quantization has become widespread for compressing image super-resolution (SR) models for edge deployment, which allows advanced SR models to enjoy compact low-bit parameters and efficient integer/bitwise constructions for storage compression and inference acceleration, respectively.

Image Super-Resolution Quantization

Binarized Diffusion Model for Image Super-Resolution

1 code implementation9 Jun 2024 Zheng Chen, Haotong Qin, Yong Guo, Xiongfei Su, Xin Yuan, Linghe Kong, Yulun Zhang

Nonetheless, due to the model structure and the multi-step iterative attribute of DMs, existing binarization methods result in significant performance degradation.

Attribute Binarization +2

Benchmarking the Robustness of Temporal Action Detection Models Against Temporal Corruptions

1 code implementation CVPR 2024 Runhao Zeng, Xiaoyong Chen, Jiaming Liang, Huisi Wu, Guangzhong Cao, Yong Guo

In this paper, we extensively analyze the robustness of seven leading TAD methods and obtain some interesting findings: 1) Existing methods are particularly vulnerable to temporal corruptions, and end-to-end methods are often more susceptible than those with a pre-trained feature extractor; 2) Vulnerability mainly comes from localization error rather than classification error; 3) When corruptions occur in the middle of an action instance, TAD models tend to yield the largest performance drop.

Action Detection Benchmarking

CustomVideo: Customizing Text-to-Video Generation with Multiple Subjects

no code implementations18 Jan 2024 Zhao Wang, Aoxue Li, Lingting Zhu, Yong Guo, Qi Dou, Zhenguo Li

Customized text-to-video generation aims to generate high-quality videos guided by text prompts and subject references.

Object Text-to-Video Generation +1

CoDe: An Explicit Content Decoupling Framework for Image Restoration

no code implementations CVPR 2024 Enxuan Gu, Hongwei Ge, Yong Guo

To address this issue we propose an explicit Content Decoupling framework for IR dubbed CoDe to end-to-end model the restoration process by utilizing decoupled content components in a divide-and-conquer-like architecture.

Image Denoising Image Restoration +1

Bridging Code Semantic and LLMs: Semantic Chain-of-Thought Prompting for Code Generation

no code implementations16 Oct 2023 Yingwei Ma, Yue Yu, Shanshan Li, Yu Jiang, Yong Guo, Yuanliang Zhang, Yutao Xie, Xiangke Liao

Meanwhile, while traditional techniques leveraging such semantic information require complex static or dynamic code analysis to obtain features such as data flow and control flow, SeCoT demonstrates that this process can be fully automated via the intrinsic capabilities of LLMs (i. e., in-context learning), while being generalizable and applicable to challenging domains.

Code Generation HumanEval +1

DriveGPT4: Interpretable End-to-end Autonomous Driving via Large Language Model

no code implementations2 Oct 2023 Zhenhua Xu, Yujia Zhang, Enze Xie, Zhen Zhao, Yong Guo, Kwan-Yee. K. Wong, Zhenguo Li, Hengshuang Zhao

Multimodal large language models (MLLMs) have emerged as a prominent area of interest within the research community, given their proficiency in handling and reasoning with non-textual data, including images and videos.

Autonomous Driving Language Modelling +2

Robustifying Token Attention for Vision Transformers

1 code implementation ICCV 2023 Yong Guo, David Stutz, Bernt Schiele

Interestingly, we observe that the attention mechanism of ViTs tends to rely on few important tokens, a phenomenon we call token overfocusing.

Semantic Segmentation

Improving Robustness of Vision Transformers by Reducing Sensitivity To Patch Corruptions

1 code implementation CVPR 2023 Yong Guo, David Stutz, Bernt Schiele

Despite their success, vision transformers still remain vulnerable to image corruptions, such as noise or blur.

Boosting Semi-Supervised Learning with Contrastive Complementary Labeling

no code implementations13 Dec 2022 Qinyi Deng, Yong Guo, Zhibang Yang, Haolin Pan, Jian Chen

In this way, these data can be also very informative if we can effectively exploit these complementary labels, i. e., the classes that a sample does not belong to.

Contrastive Learning

Downscaled Representation Matters: Improving Image Rescaling with Collaborative Downscaled Images

no code implementations ICCV 2023 Bingna Xu, Yong Guo, Luoqian Jiang, Mianjie Yu, Jian Chen

Inspired by this, we propose a Hierarchical Collaborative Downscaling (HCD) method that performs gradient descent in both HR and LR domains to improve the downscaled representations.

Image Reconstruction Super-Resolution

Pareto-aware Neural Architecture Generation for Diverse Computational Budgets

1 code implementation14 Oct 2022 Yong Guo, Yaofo Chen, Yin Zheng, Qi Chen, Peilin Zhao, Jian Chen, Junzhou Huang, Mingkui Tan

More critically, these independent search processes cannot share their learned knowledge (i. e., the distribution of good architectures) with each other and thus often result in limited search results.

Improving Fine-tuning of Self-supervised Models with Contrastive Initialization

1 code implementation30 Jul 2022 Haolin Pan, Yong Guo, Qinyi Deng, Haomin Yang, Yiqun Chen, Jian Chen

Self-supervised learning (SSL) has achieved remarkable performance in pretraining the models that can be further used in downstream tasks via fine-tuning.

Self-Supervised Learning

Towards Lightweight Super-Resolution with Dual Regression Learning

2 code implementations16 Jul 2022 Yong Guo, Mingkui Tan, Zeshuai Deng, Jingdong Wang, Qi Chen, JieZhang Cao, Yanwu Xu, Jian Chen

Nevertheless, it is hard for existing model compression methods to accurately identify the redundant components due to the extremely large SR mapping space.

Image Super-Resolution Model Compression +1

Improving Robustness by Enhancing Weak Subnets

1 code implementation30 Jan 2022 Yong Guo, David Stutz, Bernt Schiele

We show that EWS greatly improves both robustness against corrupted images as well as accuracy on clean data.

Adversarial Robustness Data Augmentation +1

AdaXpert: Adapting Neural Architecture for Growing Data

1 code implementation1 Jul 2021 Shuaicheng Niu, Jiaxiang Wu, Guanghui Xu, Yifan Zhang, Yong Guo, Peilin Zhao, Peng Wang, Mingkui Tan

To address this, we present a neural architecture adaptation method, namely Adaptation eXpert (AdaXpert), to efficiently adjust previous architectures on the growing data.

Content-Aware Convolutional Neural Networks

1 code implementation30 Jun 2021 Yong Guo, Yaofo Chen, Mingkui Tan, Kui Jia, Jian Chen, Jingdong Wang

In practice, the convolutional operation on some of the windows (e. g., smooth windows that contain very similar pixels) can be very redundant and may introduce noises into the computation.

Pareto-Frontier-aware Neural Architecture Generation for Diverse Budgets

no code implementations27 Feb 2021 Yong Guo, Yaofo Chen, Yin Zheng, Qi Chen, Peilin Zhao, Jian Chen, Junzhou Huang, Mingkui Tan

To this end, we propose a Pareto-Frontier-aware Neural Architecture Generator (NAG) which takes an arbitrary budget as input and produces the Pareto optimal architecture for the target budget.

Towards Accurate and Compact Architectures via Neural Architecture Transformer

2 code implementations20 Feb 2021 Yong Guo, Yin Zheng, Mingkui Tan, Qi Chen, Zhipeng Li, Jian Chen, Peilin Zhao, Junzhou Huang

To address this issue, we propose a Neural Architecture Transformer++ (NAT++) method which further enlarges the set of candidate transitions to improve the performance of architecture optimization.

Neural Architecture Search valid

Deep View Synthesis via Self-Consistent Generative Network

1 code implementation19 Jan 2021 Zhuoman Liu, Wei Jia, Ming Yang, Peiyao Luo, Yong Guo, Mingkui Tan

To address the above issues, in this paper, we propose a novel deep generative model, called Self-Consistent Generative Network (SCGN), which synthesizes novel views from the given input views without explicitly exploiting the geometric information.

Pareto-Frontier-aware Neural Architecture Search

no code implementations1 Jan 2021 Yong Guo, Yaofo Chen, Yin Zheng, Peilin Zhao, Jian Chen, Junzhou Huang, Mingkui Tan

To find promising architectures under different budgets, existing methods may have to perform an independent search for each budget, which is very inefficient and unnecessary.

Neural Architecture Search

Double Forward Propagation for Memorized Batch Normalization

no code implementations10 Oct 2020 Yong Guo, Qingyao Wu, Chaorui Deng, Jian Chen, Mingkui Tan

Although the standard BN can significantly accelerate the training of DNNs and improve the generalization performance, it has several underlying limitations which may hamper the performance in both training and inference.

Conditional Automated Channel Pruning for Deep Neural Networks

no code implementations21 Sep 2020 Yixin Liu, Yong Guo, Zichang Liu, Haohua Liu, Jingjie Zhang, Zejun Chen, Jing Liu, Jian Chen

To address this issue, given a target compression rate for the whole model, one can search for the optimal compression rate for each layer.

Model Compression

Improving Generative Adversarial Networks with Local Coordinate Coding

1 code implementation28 Jul 2020 Jiezhang Cao, Yong Guo, Qingyao Wu, Chunhua Shen, Junzhou Huang, Mingkui Tan

In this paper, rather than sampling from the predefined prior distribution, we propose an LCCGAN model with local coordinate coding (LCC) to improve the performance of generating data.

Breaking the Curse of Space Explosion: Towards Efficient NAS with Curriculum Search

1 code implementation ICML 2020 Yong Guo, Yaofo Chen, Yin Zheng, Peilin Zhao, Jian Chen, Junzhou Huang, Mingkui Tan

With the proposed search strategy, our Curriculum Neural Architecture Search (CNAS) method significantly improves the search efficiency and finds better architectures than existing NAS methods.

Neural Architecture Search

Disturbance-immune Weight Sharing for Neural Architecture Search

no code implementations29 Mar 2020 Shuaicheng Niu, Jiaxiang Wu, Yifan Zhang, Yong Guo, Peilin Zhao, Junzhou Huang, Mingkui Tan

To alleviate the performance disturbance issue, we propose a new disturbance-immune update strategy for model updating.

Neural Architecture Search

Hierarchical Neural Architecture Search for Single Image Super-Resolution

1 code implementation10 Mar 2020 Yong Guo, Yongsheng Luo, Zhenhao He, Jin Huang, Jian Chen

To this end, we design a hierarchical SR search space and propose a hierarchical controller for architecture search.

Image Super-Resolution Neural Architecture Search

Discrimination-aware Network Pruning for Deep Model Compression

1 code implementation4 Jan 2020 Jing Liu, Bohan Zhuang, Zhuangwei Zhuang, Yong Guo, Junzhou Huang, Jinhui Zhu, Mingkui Tan

In this paper, we propose a simple-yet-effective method called discrimination-aware channel pruning (DCP) to choose the channels that actually contribute to the discriminative power.

Face Recognition Image Classification +2

Auto-Embedding Generative Adversarial Networks for High Resolution Image Synthesis

1 code implementation27 Mar 2019 Yong Guo, Qi Chen, Jian Chen, Qingyao Wu, Qinfeng Shi, Mingkui Tan

To address this issue, we develop a novel GAN called Auto-Embedding Generative Adversarial Network (AEGAN), which simultaneously encodes the global structure features and captures the fine-grained details.

Generative Adversarial Network Image Generation +2

Learning Joint Wasserstein Auto-Encoders for Joint Distribution Matching

no code implementations27 Sep 2018 JieZhang Cao, Yong Guo, Langyuan Mo, Peilin Zhao, Junzhou Huang, Mingkui Tan

We study the joint distribution matching problem which aims at learning bidirectional mappings to match the joint distribution of two domains.

Open-Ended Question Answering Unsupervised Image-To-Image Translation +2

Dual Reconstruction Nets for Image Super-Resolution with Gradient Sensitive Loss

no code implementations19 Sep 2018 Yong Guo, Qi Chen, Jian Chen, Junzhou Huang, Yanwu Xu, JieZhang Cao, Peilin Zhao, Mingkui Tan

However, most deep learning methods employ feed-forward architectures, and thus the dependencies between LR and HR images are not fully exploited, leading to limited learning performance.

Image Super-Resolution

Adversarial Learning with Local Coordinate Coding

no code implementations ICML 2018 Jiezhang Cao, Yong Guo, Qingyao Wu, Chunhua Shen, Junzhou Huang, Mingkui Tan

Generative adversarial networks (GANs) aim to generate realistic data from some prior distribution (e. g., Gaussian noises).

The Shallow End: Empowering Shallower Deep-Convolutional Networks through Auxiliary Outputs

1 code implementation6 Nov 2016 Yong Guo, Jian Chen, Qing Du, Anton Van Den Hengel, Qinfeng Shi, Mingkui Tan

As a result, the representation power of intermediate layers can be very weak and the model becomes very redundant with limited performance.

Model Compression Model Selection

High-speed real-time single-pixel microscopy based on Fourier sampling

no code implementations15 Jun 2016 Qiang Guo, Hongwei Chen, Yuxi Wang, Yong Guo, Peng Liu, Xiurui Zhu, Zheng Cheng, Zhenming Yu, Minghua Chen, Sigang Yang, Shizhong Xie

However, according to CS theory, image reconstruction is an iterative process that consumes enormous amounts of computational time and cannot be performed in real time.

Image Reconstruction Image Restoration +1

Cannot find the paper you are looking for? You can Submit a new open access paper.