Search Results for author: Hanqing Zhu

Found 14 papers, 9 papers with code

APOLLO: SGD-like Memory, AdamW-level Performance

1 code implementation6 Dec 2024 Hanqing Zhu, Zhenyu Zhang, Wenyan Cong, Xi Liu, Sem Park, Vikas Chandra, Bo Long, David Z. Pan, Zhangyang Wang, Jinwon Lee

This memory burden necessitates using more or higher-end GPUs or reducing batch sizes, limiting training scalability and throughput.

Quantization

PACE: Pacing Operator Learning to Accurate Optical Field Simulation for Complicated Photonic Devices

1 code implementation5 Nov 2024 Hanqing Zhu, Wenyan Cong, Guojin Chen, Shupeng Ning, Ray T. Chen, Jiaqi Gu, David Z. Pan

In this work, we boost the prediction fidelity to an unprecedented level for simulating complex photonic devices with a novel operator design driven by the above challenges.

Operator learning

INSIGHT: Universal Neural Simulator for Analog Circuits Harnessing Autoregressive Transformers

no code implementations10 Jul 2024 Souradip Poddar, Youngmin Oh, Yao Lai, Hanqing Zhu, Bosun Hwang, David Z. Pan

However, efficient and effective exploration of the vast and complex design space remains constrained by the time-consuming nature of SPICE simulations, making effective design automation a challenging endeavor.

LLM-Enhanced Bayesian Optimization for Efficient Analog Layout Constraint Generation

1 code implementation7 Jun 2024 Guojin Chen, Keren Zhu, Seunggeun Kim, Hanqing Zhu, Yao Lai, Bei Yu, David Z. Pan

Analog layout synthesis faces significant challenges due to its dependence on manual processes, considerable time requirements, and performance instability.

Bayesian Optimization Few-Shot Learning

M3ICRO: Machine Learning-Enabled Compact Photonic Tensor Core based on PRogrammable Multi-Operand Multimode Interference

1 code implementation31 May 2023 Jiaqi Gu, Hanqing Zhu, Chenghao Feng, Zixuan Jiang, Ray T. Chen, David Z. Pan

The programmable MOMMI leverages the intrinsic light propagation principle, providing a single-device programmable matrix unit beyond the conventional computing paradigm of one multiply-accumulate (MAC) operation per device.

Integrated multi-operand optical neurons for scalable and hardware-efficient deep learning

no code implementations31 May 2023 Chenghao Feng, Jiaqi Gu, Hanqing Zhu, Rongxing Tang, Shupeng Ning, May Hlaing, Jason Midkiff, Sourabh Jain, David Z. Pan, Ray T. Chen

The optical neural network (ONN) is a promising hardware platform for next-generation neuromorphic computing due to its high parallelism, low latency, and low energy consumption.

Pre-RMSNorm and Pre-CRMSNorm Transformers: Equivalent and Efficient Pre-LN Transformers

1 code implementation NeurIPS 2023 Zixuan Jiang, Jiaqi Gu, Hanqing Zhu, David Z. Pan

Experiments demonstrate that we can reduce the training and inference time of Pre-LN Transformers by 1% - 10%.

NeurOLight: A Physics-Agnostic Neural Operator Enabling Parametric Photonic Device Simulation

1 code implementation19 Sep 2022 Jiaqi Gu, Zhengqi Gao, Chenghao Feng, Hanqing Zhu, Ray T. Chen, Duane S. Boning, David Z. Pan

In this work, for the first time, a physics-agnostic neural operator-based framework, dubbed NeurOLight, is proposed to learn a family of frequency-domain Maxwell PDEs for ultra-fast parametric photonic device simulation.

ELight: Enabling Efficient Photonic In-Memory Neurocomputing with Life Enhancement

no code implementations15 Dec 2021 Hanqing Zhu, Jiaqi Gu, Chenghao Feng, Mingjie Liu, Zixuan Jiang, Ray T. Chen, David Z. Pan

With the recent advances in optical phase change material (PCM), photonic in-memory neurocomputing has demonstrated its superiority in optical neural network (ONN) designs with near-zero static power consumption, time-of-light latency, and compact footprint.

A compact butterfly-style silicon photonic-electronic neural chip for hardware-efficient deep learning

1 code implementation11 Nov 2021 Chenghao Feng, Jiaqi Gu, Hanqing Zhu, Zhoufeng Ying, Zheng Zhao, David Z. Pan, Ray T. Chen

The optical neural network (ONN) is a promising hardware platform for next-generation neurocomputing due to its high parallelism, low latency, and low energy consumption.

L2ight: Enabling On-Chip Learning for Optical Neural Networks via Efficient in-situ Subspace Optimization

1 code implementation NeurIPS 2021 Jiaqi Gu, Hanqing Zhu, Chenghao Feng, Zixuan Jiang, Ray T. Chen, David Z. Pan

In this work, we propose a closed-loop ONN on-chip learning framework L2ight to enable scalable ONN mapping and efficient in-situ learning.

A Rule-Based Computational Model of Cognitive Arithmetic

no code implementations3 May 2017 Ashis Pati, Kantwon Rogers, Hanqing Zhu

This area of research explores the retrieval mechanisms and strategies used by people during a common cognitive task.

Math model +1

Cannot find the paper you are looking for? You can Submit a new open access paper.