Search Results for author: Ruizhi Qiao

Found 18 papers, 8 papers with code

Less is more: zero-shot learning from online textual documents with noise suppression

no code implementations CVPR 2016 Ruizhi Qiao, Lingqiao Liu, Chunhua Shen, Anton Van Den Hengel

Classifying a visual concept merely from its associated online textual source, such as a Wikipedia article, is an attractive research topic in zero-shot learning because it alleviates the burden of manually collecting semantic attributes.

Zero-Shot Learning

Structured Learning of Tree Potentials in CRF for Image Segmentation

no code implementations26 Mar 2017 Fayao Liu, Guosheng Lin, Ruizhi Qiao, Chunhua Shen

In this fashion, we easily achieve nonlinear learning of potential functions on both unary and pairwise terms in CRFs.

Image Segmentation Semantic Segmentation

Visually Aligned Word Embeddings for Improving Zero-shot Learning

no code implementations18 Jul 2017 Ruizhi Qiao, Lingqiao Liu, Chunhua Shen, Anton Van Den Hengel

To overcome this visual-semantic discrepancy, this work proposes an objective function to re-align the distributed word embeddings with visual information by learning a neural network to map it into a new representation called visually aligned word embedding (VAWE).

Semantic Similarity Semantic Textual Similarity +2

Contrastive Learning for Compact Single Image Dehazing

7 code implementations CVPR 2021 Haiyan Wu, Yanyun Qu, Shaohui Lin, Jian Zhou, Ruizhi Qiao, Zhizhong Zhang, Yuan Xie, Lizhuang Ma

In this paper, we propose a novel contrastive regularization (CR) built upon contrastive learning to exploit both the information of hazy images and clear images as negative and positive samples, respectively.

Contrastive Learning Image Dehazing +1

Novelty Detection via Contrastive Learning with Negative Data Augmentation

no code implementations18 Jun 2021 Chengwei Chen, Yuan Xie, Shaohui Lin, Ruizhi Qiao, Jian Zhou, Xin Tan, Yi Zhang, Lizhuang Ma

Moreover, our model is more stable for training in a non-adversarial manner, compared to other adversarial based novelty detection methods.

Clustering Contrastive Learning +4

HybridCR: Weakly-Supervised 3D Point Cloud Semantic Segmentation via Hybrid Contrastive Regularization

1 code implementation CVPR 2022 Mengtian Li, Yuan Xie, Yunhang Shen, Bo Ke, Ruizhi Qiao, Bo Ren, Shaohui Lin, Lizhuang Ma

To address the huge labeling cost in large-scale point cloud semantic segmentation, we propose a novel hybrid contrastive regularization (HybridCR) framework in weakly-supervised setting, which obtains competitive performance compared to its fully-supervised counterpart.

Semantic Segmentation Semantic Similarity +1

Open-Vocabulary Multi-Label Classification via Multi-Modal Knowledge Transfer

1 code implementation5 Jul 2022 Sunan He, Taian Guo, Tao Dai, Ruizhi Qiao, Bo Ren, Shu-Tao Xia

Specifically, our method exploits multi-modal knowledge of image-text pairs based on a vision and language pre-training (VLP) model.

Image-text matching Knowledge Distillation +7

Exploiting Feature Diversity for Make-up Temporal Video Grounding

no code implementations12 Aug 2022 Xiujun Shu, Wei Wen, Taian Guo, Sunan He, Chen Wu, Ruizhi Qiao

This technical report presents the 3rd winning solution for MTVG, a new task introduced in the 4-th Person in Context (PIC) Challenge at ACM MM 2022.

Video Grounding

See Finer, See More: Implicit Modality Alignment for Text-based Person Retrieval

1 code implementation18 Aug 2022 Xiujun Shu, Wei Wen, Haoqian Wu, Keyu Chen, Yiran Song, Ruizhi Qiao, Bo Ren, Xiao Wang

To explore the fine-grained alignment, we further propose two implicit semantic alignment paradigms: multi-level alignment (MLA) and bidirectional mask modeling (BMM).

Person Retrieval Retrieval +3

VLMAE: Vision-Language Masked Autoencoder

no code implementations19 Aug 2022 Sunan He, Taian Guo, Tao Dai, Ruizhi Qiao, Chen Wu, Xiujun Shu, Bo Ren

Image and language modeling is of crucial importance for vision-language pre-training (VLP), which aims to learn multi-modal representations from large-scale paired image-text data.

Language Modelling Question Answering +4

Collaborative Noisy Label Cleaner: Learning Scene-aware Trailers for Multi-modal Highlight Detection in Movies

1 code implementation CVPR 2023 Bei Gan, Xiujun Shu, Ruizhi Qiao, Haoqian Wu, Keyu Chen, Hanjun Li, Bo Ren

Based on existing efforts, this work has two observations: (1) For different annotators, labeling highlight has uncertainty, which leads to inaccurate and time-consuming annotations.

Highlight Detection Learning with noisy labels +1

D3G: Exploring Gaussian Prior for Temporal Sentence Grounding with Glance Annotation

1 code implementation ICCV 2023 Hanjun Li, Xiujun Shu, Sunan He, Ruizhi Qiao, Wei Wen, Taian Guo, Bei Gan, Xing Sun

Under this setup, we propose a Dynamic Gaussian prior based Grounding framework with Glance annotation (D3G), which consists of a Semantic Alignment Group Contrastive Learning module (SA-GCL) and a Dynamic Gaussian prior Adjustment module (DGA).

Contrastive Learning Sentence +1

Cannot find the paper you are looking for? You can Submit a new open access paper.