Search Results for author: Zhicai Ou

Found 5 papers, 2 papers with code

LLaVA-Phi: Efficient Multi-Modal Assistant with Small Language Model

1 code implementation4 Jan 2024 Yichen Zhu, Minjie Zhu, Ning Liu, Zhicai Ou, Xiaofeng Mou, Jian Tang

In this paper, we introduce LLaVA-$\phi$ (LLaVA-Phi), an efficient multi-modal assistant that harnesses the power of the recently advanced small language model, Phi-2, to facilitate multi-modal dialogues.

Language Modelling Visual Question Answering

ScaleKD: Distilling Scale-Aware Knowledge in Small Object Detector

no code implementations CVPR 2023 Yichen Zhu, Qiqi Zhou, Ning Liu, Zhiyuan Xu, Zhicai Ou, Xiaofeng Mou, Jian Tang

Unlike existing works that struggle to balance the trade-off between inference speed and SOD performance, in this paper, we propose a novel Scale-aware Knowledge Distillation (ScaleKD), which transfers knowledge of a complex teacher model to a compact student model.

Knowledge Distillation object-detection +2

Make A Long Image Short: Adaptive Token Length for Vision Transformers

no code implementations3 Dec 2021 Yichen Zhu, Yuqin Zhu, Jie Du, Yi Wang, Zhicai Ou, Feifei Feng, Jian Tang

The TLA enables the ReViT to process the image with the minimum sufficient number of tokens during inference.

Action Recognition Image Classification

Training BatchNorm Only in Neural Architecture Search and Beyond

no code implementations1 Dec 2021 Yichen Zhu, Jie Du, Yuqin Zhu, Yi Wang, Zhicai Ou, Feifei Feng, Jian Tang

Critically, there is no effort to understand 1) why training BatchNorm only can find the perform-well architectures with the reduced supernet-training time, and 2) what is the difference between the train-BN-only supernet and the standard-train supernet.

Fairness Neural Architecture Search

Cannot find the paper you are looking for? You can Submit a new open access paper.