Search Results for author: Xu Luo

Found 18 papers, 11 papers with code

Unleashing the Potentials of Likelihood Composition for Multi-modal Language Models

1 code implementation1 Oct 2024 Shitian Zhao, Renrui Zhang, Xu Luo, Yan Wang, Shanghang Zhang, Peng Gao

In this framework, people can propose new basic composition methods and combine them to get the new mixed composition methods.

Question Answering Visual Question Answering

Lumina-Next: Making Lumina-T2X Stronger and Faster with Next-DiT

1 code implementation5 Jun 2024 Le Zhuo, Ruoyi Du, Han Xiao, Yangguang Li, Dongyang Liu, Rongjie Huang, Wenze Liu, Lirui Zhao, Fu-Yun Wang, Zhanyu Ma, Xu Luo, Zehan Wang, Kaipeng Zhang, Xiangyang Zhu, Si Liu, Xiangyu Yue, Dingning Liu, Wanli Ouyang, Ziwei Liu, Yu Qiao, Hongsheng Li, Peng Gao

Lumina-T2X is a nascent family of Flow-based Large Diffusion Transformers that establishes a unified framework for transforming noise into various modalities, such as images and videos, conditioned on text instructions.

Point Cloud Generation Text-to-Image Generation

Lumina-T2X: Transforming Text into Any Modality, Resolution, and Duration via Flow-based Large Diffusion Transformers

2 code implementations9 May 2024 Peng Gao, Le Zhuo, Dongyang Liu, Ruoyi Du, Xu Luo, Longtian Qiu, Yuhang Zhang, Chen Lin, Rongjie Huang, Shijie Geng, Renrui Zhang, Junlin Xi, Wenqi Shao, Zhengkai Jiang, Tianshuo Yang, Weicai Ye, He Tong, Jingwen He, Yu Qiao, Hongsheng Li

Sora unveils the potential of scaling Diffusion Transformer for generating photorealistic images and videos at arbitrary resolutions, aspect ratios, and durations, yet it still lacks sufficient implementation details.

Symmetrical Bidirectional Knowledge Alignment for Zero-Shot Sketch-Based Image Retrieval

1 code implementation16 Dec 2023 Decheng Liu, Xu Luo, Chunlei Peng, Nannan Wang, Ruimin Hu, Xinbo Gao

In this paper, we propose a novel Symmetrical Bidirectional Knowledge Alignment for zero-shot sketch-based image retrieval (SBKA).

Knowledge Distillation Retrieval +1

3DAxiesPrompts: Unleashing the 3D Spatial Task Capabilities of GPT-4V

no code implementations15 Dec 2023 Dingning Liu, Xiaomeng Dong, Renrui Zhang, Xu Luo, Peng Gao, Xiaoshui Huang, Yongshun Gong, Zhihui Wang

In this work, we present a new visual prompting method called 3DAxiesPrompts (3DAP) to unleash the capabilities of GPT-4V in performing 3D spatial tasks.

3D Object Detection object-detection +1

Less is More: On the Feature Redundancy of Pretrained Models When Transferring to Few-shot Tasks

no code implementations5 Oct 2023 Xu Luo, Difan Zou, Lianli Gao, Zenglin Xu, Jingkuan Song

Transferring a pretrained model to a downstream task can be as easy as conducting linear probing with target data, that is, training a linear classifier upon frozen features extracted from the pretrained model.

Feature Importance

Enhanced Federated Optimization: Adaptive Unbiased Client Sampling with Reduced Variance

no code implementations4 Oct 2023 Dun Zeng, Zenglin Xu, Yu Pan, Xu Luo, Qifan Wang, Xiaoying Tang

Central to this process is the technique of unbiased client sampling, which ensures a representative selection of clients.

Federated Learning

Language-Enhanced Session-Based Recommendation with Decoupled Contrastive Learning

no code implementations20 Jul 2023 Zhipeng Zhang, Piao Tong, Yingwei Ma, Qiao Liu, Xujiang Liu, Xu Luo

Furthermore, we introduce a novel Decoupled Contrastive Learning method to enhance the effectiveness of the language representation.

Contrastive Learning Retrieval +1

DETA: Denoised Task Adaptation for Few-Shot Learning

3 code implementations ICCV 2023 Ji Zhang, Lianli Gao, Xu Luo, HengTao Shen, Jingkuan Song

Test-time task adaptation in few-shot learning aims to adapt a pre-trained task-agnostic model for capturing taskspecific knowledge of the test task, rely only on few-labeled support samples.

Denoising Few-Shot Learning

A Closer Look at Few-shot Classification Again

3 code implementations28 Jan 2023 Xu Luo, Hao Wu, Ji Zhang, Lianli Gao, Jing Xu, Jingkuan Song

Few-shot classification consists of a training phase where a model is learned on a relatively large dataset and an adaptation phase where the learned model is adapted to previously-unseen tasks with limited labeled samples.

Classification Representation Learning +1

Alleviating the Sample Selection Bias in Few-shot Learning by Removing Projection to the Centroid

2 code implementations30 Oct 2022 Jing Xu, Xu Luo, Xinglin Pan, Wenjie Pei, Yanan Li, Zenglin Xu

In this paper, we find that this problem usually occurs when the positions of support samples are in the vicinity of task centroid -- the mean of all class centroids in the task.

Few-Shot Learning Selection bias

Channel Importance Matters in Few-Shot Image Classification

1 code implementation16 Jun 2022 Xu Luo, Jing Xu, Zenglin Xu

When facing novel few-shot tasks in the test-time datasets, this transformation can greatly improve the generalization ability of learned image representations, while being agnostic to the choice of training algorithms and datasets.

Few-Shot Image Classification Few-Shot Learning

Exploring Category-correlated Feature for Few-shot Image Classification

no code implementations14 Dec 2021 Jing Xu, Xinglin Pan, Xu Luo, Wenjie Pei, Zenglin Xu

To alleviate this problem, we present a simple yet effective feature rectification method by exploring the category correlation between novel and base classes as the prior knowledge.

Classification Decoder +1

Adaptive Multi-layer Contrastive Graph Neural Networks

no code implementations29 Sep 2021 Shuhao Shi, Pengfei Xie, Xu Luo, Kai Qiao, Linyuan Wang, Jian Chen, Bin Yan

AMC-GNN generates two graph views by data augmentation and compares different layers' output embeddings of Graph Neural Network encoders to obtain feature representations, which could be used for downstream tasks.

Data Augmentation Graph Neural Network +1

Boosting Few-Shot Classification with View-Learnable Contrastive Learning

1 code implementation20 Jul 2021 Xu Luo, Yuxuan Chen, Liangjian Wen, Lili Pan, Zenglin Xu

The goal of few-shot classification is to classify new categories with few labeled examples within each class.

Classification Contrastive Learning +1

Cannot find the paper you are looking for? You can Submit a new open access paper.