Search Results for author: Zhiru Zhang

Found 22 papers, 8 papers with code

Structured Pruning is All You Need for Pruning CNNs at Initialization

no code implementations4 Mar 2022 Yaohui Cai, Weizhe Hua, Hongzheng Chen, G. Edward Suh, Christopher De Sa, Zhiru Zhang

In addition, since PreCropping compresses CNNs at initialization, the computational and memory costs of CNNs are reduced for both training and inference on commodity hardware.

Model Compression

Understanding Hyperdimensional Computing for Parallel Single-Pass Learning

no code implementations10 Feb 2022 Tao Yu, Yichi Zhang, Zhiru Zhang, Christopher De Sa

Using representation theory, we characterize which similarity matrices can be "expressed" by finite group VSA hypervectors, and we show how these VSAs can be constructed.

GARNET: Reduced-Rank Topology Learning for Robust and Scalable Graph Neural Networks

no code implementations30 Jan 2022 Chenhui Deng, Xiuyu Li, Zhuo Feng, Zhiru Zhang

Graph neural networks (GNNs) have been increasingly deployed in various applications that involve learning on non-Euclidean data.

Adversarial Robustness

PokeBNN: A Binary Pursuit of Lightweight Accuracy

1 code implementation30 Nov 2021 Yichi Zhang, Zhiru Zhang, Lukasz Lew

In order to enable joint optimization of the cost together with accuracy, we define arithmetic computation effort (ACE), a hardware- and energy-inspired cost metric for quantized and binarized networks.

Binarization

GARNET: A Spectral Approach to Robust and Scalable Graph Neural Networks

no code implementations29 Sep 2021 Chenhui Deng, Xiuyu Li, Zhuo Feng, Zhiru Zhang

In this paper, we propose GARNET, a scalable spectral method to boost the adversarial robustness of GNN models for both homophilic and heterophilic graphs.

Adversarial Robustness Graph Embedding

Dense Pruning of Pointwise Convolutions in the Frequency Domain

no code implementations16 Sep 2021 Mark Buckler, Neil Adit, Yuwei Hu, Zhiru Zhang, Adrian Sampson

Our key insights are that 1) pointwise convolutions commute with frequency transformation and thus can be computed in the frequency domain without modification, 2) each channel within a given layer has a different level of sensitivity to frequency domain pruning, and 3) each channel's sensitivity to frequency pruning is approximately monotonic with respect to frequency.

Enabling Design Methodologies and Future Trends for Edge AI: Specialization and Co-design

no code implementations25 Mar 2021 Cong Hao, Jordan Dotzel, JinJun Xiong, Luca Benini, Zhiru Zhang, Deming Chen

Artificial intelligence (AI) technologies have dramatically advanced in recent years, resulting in revolutionary changes in people's lives.

Edge-computing

SPADE: A Spectral Method for Black-Box Adversarial Robustness Evaluation

1 code implementation7 Feb 2021 Wuxinlin Cheng, Chenhui Deng, Zhiqiang Zhao, Yaohui Cai, Zhiru Zhang, Zhuo Feng

A black-box spectral method is introduced for evaluating the adversarial robustness of a given machine learning (ML) model.

Adversarial Robustness Graph Embedding

GuardNN: Secure DNN Accelerator for Privacy-Preserving Deep Learning

no code implementations26 Aug 2020 Weizhe Hua, Muhammad Umar, Zhiru Zhang, G. Edward Suh

GuardNN shows that the architecture and protection can be customized for a specific application to provide strong confidentiality and integrity protection with negligible overhead.

Privacy Preserving Deep Learning

FeatGraph: A Flexible and Efficient Backend for Graph Neural Network Systems

no code implementations26 Aug 2020 Yuwei Hu, Zihao Ye, Minjie Wang, Jiali Yu, Da Zheng, Mu Li, Zheng Zhang, Zhiru Zhang, Yida Wang

FeatGraph provides a flexible programming interface to express diverse GNN models by composing coarse-grained sparse templates with fine-grained user-defined functions (UDFs) on each vertex/edge.

MgX: Near-Zero Overhead Memory Protection with an Application to Secure DNN Acceleration

no code implementations20 Apr 2020 Weizhe Hua, Muhammad Umar, Zhiru Zhang, G. Edward Suh

In this paper, we propose MgX, a near-zero overhead memory protection scheme for hardware accelerators.

Precision Gating: Improving Neural Network Efficiency with Dynamic Dual-Precision Activations

1 code implementation ICLR 2020 Yichi Zhang, Ritchie Zhao, Weizhe Hua, Nayun Xu, G. Edward Suh, Zhiru Zhang

The proposed approach is applicable to a variety of DNN architectures and significantly reduces the computational cost of DNN execution with almost no accuracy loss.

Quantization

OverQ: Opportunistic Outlier Quantization for Neural Network Accelerators

no code implementations13 Oct 2019 Ritchie Zhao, Jordan Dotzel, Zhanqiu Hu, Preslav Ivanov, Christopher De Sa, Zhiru Zhang

Specialized hardware for handling activation outliers can enable low-precision neural networks, but at the cost of nontrivial area overhead.

Quantization

GraphZoom: A multi-level spectral approach for accurate and scalable graph embedding

1 code implementation ICLR 2020 Chenhui Deng, Zhiqiang Zhao, Yongyu Wang, Zhiru Zhang, Zhuo Feng

GraphZoom first performs graph fusion to generate a new graph that effectively encodes the topology of the original graph and the node attribute information.

Graph Embedding

Painting on Placement: Forecasting Routing Congestion using Conditional Generative Adversarial Nets

no code implementations15 Apr 2019 Cunxi Yu, Zhiru Zhang

Physical design process commonly consumes hours to days for large designs, and routing is known as the most critical step.

Colorization Translation

Improving Neural Network Quantization without Retraining using Outlier Channel Splitting

3 code implementations28 Jan 2019 Ritchie Zhao, Yuwei Hu, Jordan Dotzel, Christopher De Sa, Zhiru Zhang

The majority of existing literature focuses on training quantized DNNs, while this work examines the less-studied topic of quantizing a floating-point model without (re)training.

Language Modelling Neural Network Compression +1

Building Efficient Deep Neural Networks with Unitary Group Convolutions

no code implementations CVPR 2019 Ritchie Zhao, Yuwei Hu, Jordan Dotzel, Christopher De Sa, Zhiru Zhang

UGConvs generalize two disparate ideas in CNN architecture, channel shuffling (i. e. ShuffleNet) and block-circulant networks (i. e. CirCNN), and provide unifying insights that lead to a deeper understanding of each technique.

Channel Gating Neural Networks

1 code implementation NeurIPS 2019 Weizhe Hua, Yuan Zhou, Christopher De Sa, Zhiru Zhang, G. Edward Suh

Combining our method with knowledge distillation reduces the compute cost of ResNet-18 by 2. 6$\times$ without accuracy drop on ImageNet.

Knowledge Distillation Network Pruning

Binarized Convolutional Neural Networks with Separable Filters for Efficient Hardware Acceleration

no code implementations15 Jul 2017 Jeng-Hau Lin, Tianwei Xing, Ritchie Zhao, Zhiru Zhang, Mani Srivastava, Zhuowen Tu, Rajesh K. Gupta

State-of-the-art convolutional neural networks are enormously costly in both compute and memory, demanding massively parallel GPUs for execution.

Cannot find the paper you are looking for? You can Submit a new open access paper.