Search Results for author: Yuting Gao

Found 22 papers, 10 papers with code

Multi-Modal Prompt Learning on Blind Image Quality Assessment

1 code implementation23 Apr 2024 Wensheng Pan, Timin Gao, Yan Zhang, Runze Hu, Xiawu Zheng, Enwei Zhang, Yuting Gao, Yutao Liu, Yunhang Shen, Ke Li, Shengchuan Zhang, Liujuan Cao, Rongrong Ji

Image Quality Assessment (IQA) models benefit significantly from semantic information, which allows them to treat different types of objects distinctly.

RESTORE: Towards Feature Shift for Vision-Language Prompt Learning

1 code implementation10 Mar 2024 Yuncheng Yang, Chuyan Zhang, Zuopeng Yang, Yuting Gao, Yulei Qin, Ke Li, Xing Sun, Jie Yang, Yun Gu

Prompt learning is effective for fine-tuning foundation models to improve their generalization across a variety of downstream tasks.

Sinkhorn Distance Minimization for Knowledge Distillation

1 code implementation27 Feb 2024 Xiao Cui, Yulei Qin, Yuting Gao, Enwei Zhang, Zihan Xu, Tong Wu, Ke Li, Xing Sun, Wengang Zhou, Houqiang Li

We propose the Sinkhorn Knowledge Distillation (SinKD) that exploits the Sinkhorn distance to ensure a nuanced and precise assessment of the disparity between teacher and student distributions.

Decoder Knowledge Distillation

MMICT: Boosting Multi-Modal Fine-Tuning with In-Context Examples

1 code implementation11 Dec 2023 Tao Chen, Enwei Zhang, Yuting Gao, Ke Li, Xing Sun, Yan Zhang, Hui Li, Rongrong Ji

Although In-Context Learning (ICL) brings remarkable performance gains to Large Language Models (LLMs), the improvements remain lower than fine-tuning on downstream tasks.

In-Context Learning

Less is More: Learning Reference Knowledge Using No-Reference Image Quality Assessment

no code implementations1 Dec 2023 Xudong Li, Jingyuan Zheng, Xiawu Zheng, Runze Hu, Enwei Zhang, Yuting Gao, Yunhang Shen, Ke Li, Yutao Liu, Pingyang Dai, Yan Zhang, Rongrong Ji

Concretely, by innovatively introducing a novel feature distillation method in IQA, we propose a new framework to learn comparative knowledge from non-aligned reference images.

Inductive Bias NR-IQA

SoftCLIP: Softer Cross-modal Alignment Makes CLIP Stronger

no code implementations30 Mar 2023 Yuting Gao, Jinfeng Liu, Zihan Xu, Tong Wu Enwei Zhang, Wei Liu, Jie Yang, Ke Li, Xing Sun

During the preceding biennium, vision-language pre-training has achieved noteworthy success on several downstream tasks.

cross-modal alignment Zero-Shot Learning

Efficient Decoder-free Object Detection with Transformers

2 code implementations14 Jun 2022 Peixian Chen, Mengdan Zhang, Yunhang Shen, Kekai Sheng, Yuting Gao, Xing Sun, Ke Li, Chunhua Shen

A natural usage of ViTs in detection is to replace the CNN-based backbone with a transformer-based backbone, which is straightforward and effective, with the price of bringing considerable computation burden for inference.

Decoder Object +1

Self-supervised Models are Good Teaching Assistants for Vision Transformers

no code implementations29 Sep 2021 Haiyan Wu, Yuting Gao, Ke Li, Yinqi Zhang, Shaohui Lin, Yuan Xie, Xing Sun

These findings motivate us to introduce an self-supervised teaching assistant (SSTA) besides the commonly used supervised teacher to improve the performance of transformers.

Image Classification Knowledge Distillation

DisCo: Remedy Self-supervised Learning on Lightweight Models with Distilled Contrastive Learning

2 code implementations19 Apr 2021 Yuting Gao, Jia-Xin Zhuang, Shaohui Lin, Hao Cheng, Xing Sun, Ke Li, Chunhua Shen

Specifically, we find the final embedding obtained by the mainstream SSL methods contains the most fruitful information, and propose to distill the final embedding to maximally transmit a teacher's knowledge to a lightweight model by constraining the last embedding of the student to be consistent with that of the teacher.

Contrastive Learning Representation Learning +1

High-throughput fast full-color digital pathology based on Fourier ptychographic microscopy via color transfer

no code implementations19 Jan 2021 Yuting Gao, Jiurun Chen, Aiye Wang, An Pan, Caiwen Ma, Baoli Yao

However, the conventional full-color digital pathology based on FPM is still time-consuming due to the repeated experiments with tri-wavelengths.

Colorization

An Empirical Study and Analysis on Open-Set Semi-Supervised Learning

no code implementations19 Jan 2021 Huixiang Luo, Hao Cheng, Fanxu Meng, Yuting Gao, Ke Li, Mengdan Zhang, Xing Sun

Pseudo-labeling (PL) and Data Augmentation-based Consistency Training (DACT) are two approaches widely used in Semi-Supervised Learning (SSL) methods.

Data Augmentation

Generative and Discriminative Learning for Distorted Image Restoration

no code implementations11 Nov 2020 Yi Gu, Yuting Gao, Jie Li, Chentao Wu, Weijia Jia

Due to the uncertainty in the distortion variation, restoring distorted images caused by liquify filter is a challenging task.

Image Restoration

Removing the Background by Adding the Background: Towards Background Robust Self-supervised Video Representation Learning

2 code implementations CVPR 2021 Jinpeng Wang, Yuting Gao, Ke Li, Yiqi Lin, Andy J. Ma, Hao Cheng, Pai Peng, Feiyue Huang, Rongrong Ji, Xing Sun

Then we force the model to pull the feature of the distracting video and the feature of the original video closer, so that the model is explicitly restricted to resist the background influence, focusing more on the motion changes.

Representation Learning Self-Supervised Learning

Edge effect removal in Fourier ptychographic microscopy via periodic plus smooth image decomposition

no code implementations7 Sep 2020 An Pan, Aiye Wang, Junfu Zheng, Yuting Gao, Caiwen Ma, Baoli Yao

Fourier ptychographic microscopy (FPM) is a promising computational imaging technique with high resolution, wide field-of-view (FOV) and quantitative phase recovery.

Automatic Remaining Useful Life Estimation Framework with Embedded Convolutional LSTM as the Backbone

no code implementations10 Aug 2020 Yexu Zhou, Yuting Gao, Yiran Huang, Michael Hefenbrock, Till Riedel, Michael Beigl

An essential task in predictive maintenance is the prediction of the Remaining Useful Life (RUL) through the analysis of multivariate time series.

Bayesian Optimization Time Series +1

Filter Grafting for Deep Neural Networks: Reason, Method, and Cultivation

1 code implementation26 Apr 2020 Hao Cheng, Fanxu Meng, Ke Li, Yuting Gao, Guangming Lu, Xing Sun, Rongrong Ji

To gain a universal improvement on both valid and invalid filters, we compensate grafting with distillation (\textbf{Cultivation}) to overcome the drawback of grafting .

valid

Double Supervised Network with Attention Mechanism for Scene Text Recognition

no code implementations2 Aug 2018 Yuting Gao, Zheng Huang, Yuchen Dai, Cheng Xu, Kai Chen, Jie Tuo

In this paper, we propose Double Supervised Network with Attention Mechanism (DSAN), a novel end-to-end trainable framework for scene text recognition.

Scene Text Recognition

Fused Text Segmentation Networks for Multi-oriented Scene Text Detection

no code implementations11 Sep 2017 Yuchen Dai, Zheng Huang, Yuting Gao, Youxuan Xu, Kai Chen, Jie Guo, Weidong Qiu

In this paper, we introduce a novel end-end framework for multi-oriented scene text detection from an instance-aware semantic segmentation perspective.

Multi-Oriented Scene Text Detection object-detection +6

Cannot find the paper you are looking for? You can Submit a new open access paper.