Search Results for author: Yunke Wang

Found 6 papers, 1 papers with code

Collage Prompting: Budget-Friendly Visual Recognition with GPT-4V

no code implementations18 Mar 2024 Siyu Xu, Yunke Wang, Daochang Liu, Chang Xu

Based on the observation that the accuracy of GPT-4V's image recognition varies significantly with the order of images within the collage prompt, our method further learns to optimize the arrangement of images for maximum recognition accuracy.

Navigate

Visual Imitation Learning with Calibrated Contrastive Representation

no code implementations21 Jan 2024 Yunke Wang, Linwei Tao, Bo Du, Yutian Lin, Chang Xu

Adversarial Imitation Learning (AIL) allows the agent to reproduce expert behavior with low-dimensional states and actions.

Contrastive Learning Imitation Learning

MAEDiff: Masked Autoencoder-enhanced Diffusion Models for Unsupervised Anomaly Detection in Brain Images

no code implementations19 Jan 2024 Rui Xu, Yunke Wang, Bo Du

To address these two issues, we propose a novel Masked Autoencoder-enhanced Diffusion Model (MAEDiff) for unsupervised anomaly detection in brain images.

Unsupervised Anomaly Detection

Imitation Learning from Purified Demonstration

no code implementations11 Oct 2023 Yunke Wang, Minjing Dong, Bo Du, Chang Xu

To tackle these problems, we propose to purify the potential perturbations in imperfect demonstrations and subsequently conduct imitation learning from purified demonstrations.

Imitation Learning

Unlabeled Imperfect Demonstrations in Adversarial Imitation Learning

1 code implementation13 Feb 2023 Yunke Wang, Bo Du, Chang Xu

The trajectories of an initial agent policy could be closer to those non-optimal expert demonstrations, but within the framework of adversarial imitation learning, agent policy will be optimized to cheat the discriminator and produce trajectories that are similar to those optimal expert demonstrations.

Imitation Learning

Multi-Tailed Vision Transformer for Efficient Inference

no code implementations3 Mar 2022 Yunke Wang, Bo Du, Wenyuan Wang, Chang Xu

To satisfy the sequential input of Transformer, the tail of ViT first splits each image into a sequence of visual tokens with a fixed length.

Cannot find the paper you are looking for? You can Submit a new open access paper.