Search Results for author: Zhemin Zhang

Found 9 papers, 1 papers with code

Generating Multi-Center Classifier via Conditional Gaussian Distribution

1 code implementation29 Jan 2024 Zhemin Zhang, Xun Gong

Specifically, we create a conditional Gaussian distribution for each class and then sample multiple sub-centers from that distribution to extend the linear classifier.

Image Classification

Vision Big Bird: Random Sparsification for Full Attention

no code implementations10 Nov 2023 Zhemin Zhang, Xun Gong

Inspired by one of the most successful transformers-based models for NLP: Big Bird, we propose a novel sparse attention mechanism for Vision Transformers (ViT).

Axially Expanded Windows for Local-Global Interaction in Vision Transformers

no code implementations19 Sep 2022 Zhemin Zhang, Xun Gong

Recently, Transformers have shown promising performance in various vision tasks.

Self-Supervised Implicit Attention: Guided Attention by The Model Itself

no code implementations15 Jun 2022 Jinyi Wu, Xun Gong, Zhemin Zhang

To verify the effectiveness of SSIA, we performed a particular implementation (called an SSIA block) in convolutional neural network models and validated it on several image classification datasets.

Image Classification Self-Supervised Learning

Positional Label for Self-Supervised Vision Transformer

no code implementations10 Jun 2022 Zhemin Zhang, Xun Gong

Positional encoding is important for vision transformer (ViT) to capture the spatial structure of the input image.

Position

ReplaceBlock: An improved regularization method based on background information

no code implementations30 Mar 2022 Zhemin Zhang, Xun Gong, Jinyi Wu

In this way, ReplaceBlock can effectively simulate the feature map of the occluded image.

Object

The Fixed Sub-Center: A Better Way to Capture Data Complexity

no code implementations24 Mar 2022 Zhemin Zhang, Xun Gong

The F-SC specifically, first samples a class center Ui for each class from a uniform distribution, and then generates a normal distribution for each class, where the mean is equal to Ui.

Image Classification

Layerwise Recurrent Autoencoder for General Real-world Traffic Flow Forecasting

no code implementations27 Sep 2018 Peize Zhao, Danfeng Cai, Shaokun Zhang, Feng Chen, Zhemin Zhang, Cheng Wang, Jonathan Li

To forecast the traffic flow across a wide area and overcome the mentioned challenges, we design and propose a promising forecasting model called Layerwise Recurrent Autoencoder (LRA), in which a three-layer stacked autoencoder (SAE) architecture is used to obtain temporal traffic correlations and a recurrent neural networks (RNNs) model for prediction.

Management

Cannot find the paper you are looking for? You can Submit a new open access paper.