Search Results for author: Yueyao Yu

Found 7 papers, 1 papers with code

Why "classic" Transformers are shallow and how to make them go deep

no code implementations11 Dec 2023 Yueyao Yu, Yin Zhang

Since its introduction in 2017, Transformer has emerged as the leading neural network architecture, catalyzing revolutionary advancements in many AI disciplines.

POViT: Vision Transformer for Multi-objective Design and Characterization of Nanophotonic Devices

no code implementations17 May 2022 Xinyu Chen, Renjie Li, Yueyao Yu, Yuanwen Shen, Wenye Li, Zhaoyu Zhang, Yin Zhang

In this work, we propose the first-ever Transformer model (POViT) to efficiently design and simulate semiconductor photonic devices with multiple objectives.

Variability of Neural Networks and Han-Layer: A Variability-Inspired Model

no code implementations29 Sep 2021 Yueyao Yu, Yin Zhang

We introduce a notion of variability to view such issues under the setting of a fixed number of parameters which is, in general, a dominant cost-factor.

A Lightweight and Gradient-Stable Neural Layer

1 code implementation8 Jun 2021 Yueyao Yu, Yin Zhang

To enhance resource efficiency and model deployability of neural networks, we propose a neural-layer architecture based on Householder weighting and absolute-value activating, called Householder-absolute neural layer or simply Han-layer.

Vocal Bursts Intensity Prediction

Multi-layer Perceptron Trainability Explained via Variability

no code implementations19 May 2021 Yueyao Yu, Yin Zhang

Despite the tremendous successes of deep neural networks (DNNs) in various applications, many fundamental aspects of deep learning remain incompletely understood, including DNN trainability.

The simpler the better: vanilla sgd revisited

no code implementations1 Jan 2021 Yueyao Yu, Jie Wang, Wenye Li, Yin Zhang

The stochastic gradient descent (SGD) method, first proposed in 1950's, has been the foundation for deep-neural-network (DNN) training with numerous enhancements including adding a momentum or adaptively selecting learning rates, or using both strategies and more.

Image Classification speech-recognition +1

AuxBlocks: Defense Adversarial Example via Auxiliary Blocks

no code implementations18 Feb 2019 Yueyao Yu, Pengfei Yu, Wenye Li

Deep learning models are vulnerable to adversarial examples, which poses an indisputable threat to their applications.

Cannot find the paper you are looking for? You can Submit a new open access paper.