Search Results for author: Houwen Peng

Found 32 papers, 26 papers with code

Common 7B Language Models Already Possess Strong Math Capabilities

no code implementations7 Mar 2024 Chen Li, Weiqi Wang, Jingcheng Hu, Yixuan Wei, Nanning Zheng, Han Hu, Zheng Zhang, Houwen Peng

This paper shows that the LLaMA-2 7B model with common pre-training already exhibits strong mathematical abilities, as evidenced by its impressive accuracy of 97. 7% and 72. 0% on the GSM8K and MATH benchmarks, respectively, when selecting the best response from 256 random generations.

GSM8K Math

TinyCLIP: CLIP Distillation via Affinity Mimicking and Weight Inheritance

1 code implementation ICCV 2023 Kan Wu, Houwen Peng, Zhenghong Zhou, Bin Xiao, Mengchen Liu, Lu Yuan, Hong Xuan, Michael Valenzuela, Xi, Chen, Xinggang Wang, Hongyang Chao, Han Hu

In this paper, we propose a novel cross-modal distillation method, called TinyCLIP, for large-scale language-image pre-trained models.

EfficientViT: Memory Efficient Vision Transformer with Cascaded Group Attention

3 code implementations CVPR 2023 Xinyu Liu, Houwen Peng, Ningxin Zheng, Yuqing Yang, Han Hu, Yixuan Yuan

Comprehensive experiments demonstrate EfficientViT outperforms existing efficient models, striking a good trade-off between speed and accuracy.

iCLIP: Bridging Image Classification and Contrastive Language-Image Pre-Training for Visual Recognition

no code implementations CVPR 2023 Yixuan Wei, Yue Cao, Zheng Zhang, Houwen Peng, Zhuliang Yao, Zhenda Xie, Han Hu, Baining Guo

This paper presents a method that effectively combines two prevalent visual recognition methods, i. e., image classification and contrastive language-image pre-training, dubbed iCLIP.

Classification Image Classification +2

Attentive Mask CLIP

1 code implementation ICCV 2023 Yifan Yang, Weiquan Huang, Yixuan Wei, Houwen Peng, Xinyang Jiang, Huiqiang Jiang, Fangyun Wei, Yin Wang, Han Hu, Lili Qiu, Yuqing Yang

To address this issue, we propose an attentive token removal approach for CLIP training, which retains tokens with a high semantic correlation to the text description.

Contrastive Learning Retrieval +1

TinyViT: Fast Pretraining Distillation for Small Vision Transformers

2 code implementations21 Jul 2022 Kan Wu, Jinnian Zhang, Houwen Peng, Mengchen Liu, Bin Xiao, Jianlong Fu, Lu Yuan

It achieves a top-1 accuracy of 84. 8% on ImageNet-1k with only 21M parameters, being comparable to Swin-B pretrained on ImageNet-21k while using 4. 2 times fewer parameters.

Image Classification Knowledge Distillation

Searching the Search Space of Vision Transformer

2 code implementations NeurIPS 2021 Minghao Chen, Kan Wu, Bolin Ni, Houwen Peng, Bei Liu, Jianlong Fu, Hongyang Chao, Haibin Ling

Vision Transformer has shown great visual representation power in substantial vision tasks such as recognition and detection, and thus been attracting fast-growing efforts on manually designing more effective architectures.

Neural Architecture Search object-detection +4

AutoFormer: Searching Transformers for Visual Recognition

2 code implementations ICCV 2021 Minghao Chen, Houwen Peng, Jianlong Fu, Haibin Ling

Specifically, the performance of these subnets with weights inherited from the supernet is comparable to those retrained from scratch.

AutoML Fine-Grained Image Classification

One-Shot Neural Ensemble Architecture Search by Diversity-Guided Search Space Shrinking

1 code implementation CVPR 2021 Minghao Chen, Houwen Peng, Jianlong Fu, Haibin Ling

In this paper, we propose a one-shot neural ensemble architecture search (NEAS) solution that addresses the two challenges.

Neural Architecture Search

Multi-Scale 2D Temporal Adjacent Networks for Moment Localization with Natural Language

1 code implementation4 Dec 2020 Songyang Zhang, Houwen Peng, Jianlong Fu, Yijuan Lu, Jiebo Luo

It is a challenging problem because a target moment may take place in the context of other temporal moments in the untrimmed video.

Cream of the Crop: Distilling Prioritized Paths For One-Shot Neural Architecture Search

2 code implementations NeurIPS 2020 Houwen Peng, Hao Du, Hongyuan Yu, Qi Li, Jing Liao, Jianlong Fu

The experiments on ImageNet verify such path distillation method can improve the convergence ratio and performance of the hypernetwork, as well as boosting the training of subnetworks.

Neural Architecture Search object-detection +1

Revisiting Anchor Mechanisms for Temporal Action Localization

1 code implementation22 Aug 2020 Le Yang, Houwen Peng, Dingwen Zhang, Jianlong Fu, Junwei Han

To address this problem, this paper proposes a novel anchor-free action localization module that assists action localization by temporal points.

Temporal Action Localization

Towards Accurate Pixel-wise Object Tracking by Attention Retrieval

1 code implementation6 Aug 2020 Zhipeng Zhang, Bing Li, Weiming Hu, Houwen Peng

We first build a look-up-table (LUT) with the ground-truth mask in the starting frame, and then retrieves the LUT to obtain an attention map for spatial constraints.

Object Object Tracking +2

Cyclic Differentiable Architecture Search

3 code implementations18 Jun 2020 Hongyuan Yu, Houwen Peng, Yan Huang, Jianlong Fu, Hao Du, Liang Wang, Haibin Ling

First, the search network generates an initial architecture for evaluation, and the weights of the evaluation network are optimized.

Neural Architecture Search

Learning 2D Temporal Adjacent Networks for Moment Localization with Natural Language

3 code implementations8 Dec 2019 Songyang Zhang, Houwen Peng, Jianlong Fu, Jiebo Luo

We address the problem of retrieving a specific moment from an untrimmed video by a query sentence.

Sentence

Learning Sparse 2D Temporal Adjacent Networks for Temporal Action Localization

2 code implementations8 Dec 2019 Songyang Zhang, Houwen Peng, Le Yang, Jianlong Fu, Jiebo Luo

In this report, we introduce the Winner method for HACS Temporal Action Localization Challenge 2019.

Temporal Action Localization

Illumination Estimation Based on Bilayer Sparse Coding

no code implementations CVPR 2013 Bing Li, Weihua Xiong, Weiming Hu, Houwen Peng

In this paper, we propose a novel bilayer sparse coding model for illumination estimation that considers image similarity in terms of both low level color distribution and high level image scene content simultaneously.

Color Constancy

Cannot find the paper you are looking for? You can Submit a new open access paper.