Search Results for author: Aoxue Li

Found 19 papers, 5 papers with code

Efficient Transferability Assessment for Selection of Pre-trained Detectors

no code implementations14 Mar 2024 Zhao Wang, Aoxue Li, Zhenguo Li, Qi Dou

Given this zoo, we adopt 7 target datasets from 5 diverse domains as the downstream target tasks for evaluation.

Open-Vocabulary Object Detection with Meta Prompt Representation and Instance Contrastive Optimization

no code implementations14 Mar 2024 Zhao Wang, Aoxue Li, Fengwei Zhou, Zhenguo Li, Qi Dou

Without using knowledge distillation, ensemble model or extra training data during detector training, our proposed MIC outperforms previous SOTA methods trained with these complex techniques on LVIS.

Contrastive Learning Knowledge Distillation +2

Divide and Conquer: Language Models can Plan and Self-Correct for Compositional Text-to-Image Generation

no code implementations28 Jan 2024 Zhenyu Wang, Enze Xie, Aoxue Li, Zhongdao Wang, Xihui Liu, Zhenguo Li

Given a complex text prompt containing multiple concepts including objects, attributes, and relationships, the LLM agent initially decomposes it, which entails the extraction of individual objects, their associated attributes, and the prediction of a coherent scene layout.

Attribute Language Modelling +3

CustomVideo: Customizing Text-to-Video Generation with Multiple Subjects

no code implementations18 Jan 2024 Zhao Wang, Aoxue Li, Enze Xie, Lingting Zhu, Yong Guo, Qi Dou, Zhenguo Li

Customized text-to-video generation aims to generate high-quality videos guided by text prompts and subject references.

Object Text-to-Video Generation +1

Mixture of Cluster-conditional LoRA Experts for Vision-language Instruction Tuning

no code implementations19 Dec 2023 Yunhao Gou, Zhili Liu, Kai Chen, Lanqing Hong, Hang Xu, Aoxue Li, Dit-yan Yeung, James T. Kwok, Yu Zhang

Instruction tuning of Large Vision-language Models (LVLMs) has revolutionized the development of versatile models with zero-shot generalization across a wide range of downstream vision-language tasks.

Instruction Following Zero-shot Generalization

ContraNeRF: Generalizable Neural Radiance Fields for Synthetic-to-real Novel View Synthesis via Contrastive Learning

no code implementations CVPR 2023 Hao Yang, Lanqing Hong, Aoxue Li, Tianyang Hu, Zhenguo Li, Gim Hee Lee, LiWei Wang

In this work, we first investigate the effects of synthetic data in synthetic-to-real novel view synthesis and surprisingly observe that models trained with synthetic data tend to produce sharper but less accurate volume densities.

Contrastive Learning Generalizable Novel View Synthesis +2

STC-IDS: Spatial-Temporal Correlation Feature Analyzing based Intrusion Detection System for Intelligent Connected Vehicles

no code implementations23 Apr 2022 Pengzhou Cheng, Mu Han, Aoxue Li, Fengwei Zhang

To address these limitations, we present a novel model for automotive intrusion detection by spatial-temporal correlation features of in-vehicle communication traffic (STC-IDS).

Anomaly Classification Bayesian Optimization +3

Semi-Supervised Object Detection via Multi-Instance Alignment With Global Class Prototypes

no code implementations CVPR 2022 Aoxue Li, Peng Yuan, Zhenguo Li

Semi-Supervised object detection (SSOD) aims to improve the generalization ability of object detectors with large-scale unlabeled images.

object-detection Object Detection +1

Transformation Invariant Few-Shot Object Detection

no code implementations CVPR 2021 Aoxue Li, Zhenguo Li

To this end, we propose a simple yet effective Transformation Invariant Principle (TIP) that can be flexibly applied to various meta-learning models for boosting the detection performance on novel class objects.

Few-Shot Object Detection Meta-Learning +2

Dense Relation Distillation with Context-aware Aggregation for Few-Shot Object Detection

1 code implementation CVPR 2021 Hanzhe Hu, Shuai Bai, Aoxue Li, Jinshi Cui, LiWei Wang

In this work, aiming to fully exploit features of annotated novel object and capture fine-grained features of query object, we propose Dense Relation Distillation with Context-aware Aggregation (DCNet) to tackle the few-shot detection problem.

Few-Shot Object Detection Meta-Learning +3

Boosting Few-Shot Learning With Adaptive Margin Loss

no code implementations CVPR 2020 Aoxue Li, Weiran Huang, Xu Lan, Jiashi Feng, Zhenguo Li, Li-Wei Wang

Few-shot learning (FSL) has attracted increasing attention in recent years but remains challenging, due to the intrinsic difficulty in learning to generalize from a few examples.

Few-Shot Image Classification Few-Shot Learning +2

Few-Shot Learning with Global Class Representations

2 code implementations ICCV 2019 Tiange Luo, Aoxue Li, Tao Xiang, Weiran Huang, Li-Wei Wang

In this paper, we propose to tackle the challenging few-shot learning (FSL) problem by learning global class representations using both base and novel class training samples.

Few-Shot Learning Generalized Few-Shot Classification

Transferrable Feature and Projection Learning with Class Hierarchy for Zero-Shot Learning

no code implementations19 Oct 2018 Aoxue Li, Zhiwu Lu, Jiechao Guan, Tao Xiang, Li-Wei Wang, Ji-Rong Wen

Inspired by the fact that an unseen class is not exactly `unseen' if it belongs to the same superclass as a seen class, we propose a novel inductive ZSL model that leverages superclasses as the bridge between seen and unseen classes to narrow the domain gap.

Attribute Clustering +2

Zero and Few Shot Learning with Semantic Feature Synthesis and Competitive Learning

no code implementations19 Oct 2018 Zhiwu Lu, Jiechao Guan, Aoxue Li, Tao Xiang, An Zhao, Ji-Rong Wen

Specifically, we assume that each synthesised data point can belong to any unseen class; and the most likely two class candidates are exploited to learn a robust projection function in a competitive fashion.

Attribute Few-Shot Learning +2

Zero-Shot Fine-Grained Classification by Deep Feature Learning with Semantics

no code implementations4 Jul 2017 Aoxue Li, Zhiwu Lu, Li-Wei Wang, Tao Xiang, Xinqi Li, Ji-Rong Wen

In this paper, to address the two issues, we propose a two-phase framework for recognizing images from unseen fine-grained classes, i. e. zero-shot fine-grained classification.

Classification Domain Adaptation +3

Cannot find the paper you are looking for? You can Submit a new open access paper.