Search Results for author: G. Edward Suh

Found 12 papers, 3 papers with code

Approximating ReLU on a Reduced Ring for Efficient MPC-based Private Inference

no code implementations9 Sep 2023 Kiwan Maeng, G. Edward Suh

Secure multi-party computation (MPC) allows users to offload machine learning inference on untrusted servers without having to share their privacy-sensitive data.

GPU-based Private Information Retrieval for On-Device Machine Learning Inference

1 code implementation26 Jan 2023 Maximilian Lam, Jeff Johnson, Wenjie Xiong, Kiwan Maeng, Udit Gupta, Yang Li, Liangzhen Lai, Ilias Leontiadis, Minsoo Rhu, Hsien-Hsin S. Lee, Vijay Janapa Reddi, Gu-Yeon Wei, David Brooks, G. Edward Suh

Together, for various on-device ML applications such as recommendation and language modeling, our system on a single V100 GPU can serve up to $100, 000$ queries per second -- a $>100 \times$ throughput improvement over a CPU-based baseline -- while maintaining model accuracy.

Information Retrieval Language Modelling +1

Data Leakage via Access Patterns of Sparse Features in Deep Learning-based Recommendation Systems

no code implementations12 Dec 2022 Hanieh Hashemi, Wenjie Xiong, Liu Ke, Kiwan Maeng, Murali Annavaram, G. Edward Suh, Hsien-Hsin S. Lee

This paper explores the private information that may be learned by tracking a recommendation model's sparse feature access patterns.

Recommendation Systems

Structured Pruning is All You Need for Pruning CNNs at Initialization

no code implementations4 Mar 2022 Yaohui Cai, Weizhe Hua, Hongzheng Chen, G. Edward Suh, Christopher De Sa, Zhiru Zhang

In addition, since PreCropping compresses CNNs at initialization, the computational and memory costs of CNNs are reduced for both training and inference on commodity hardware.

Model Compression

GuardNN: Secure Accelerator Architecture for Privacy-Preserving Deep Learning

no code implementations26 Aug 2020 Weizhe Hua, Muhammad Umar, Zhiru Zhang, G. Edward Suh

This paper proposes GuardNN, a secure DNN accelerator that provides hardware-based protection for user data and model parameters even in an untrusted environment.

Privacy Preserving Privacy Preserving Deep Learning

MGX: Near-Zero Overhead Memory Protection for Data-Intensive Accelerators

no code implementations20 Apr 2020 Weizhe Hua, Muhammad Umar, Zhiru Zhang, G. Edward Suh

This paper introduces MGX, a near-zero overhead memory protection scheme for hardware accelerators.

Precision Gating: Improving Neural Network Efficiency with Dynamic Dual-Precision Activations

1 code implementation ICLR 2020 Yichi Zhang, Ritchie Zhao, Weizhe Hua, Nayun Xu, G. Edward Suh, Zhiru Zhang

The proposed approach is applicable to a variety of DNN architectures and significantly reduces the computational cost of DNN execution with almost no accuracy loss.

Quantization

Channel Gating Neural Networks

1 code implementation NeurIPS 2019 Weizhe Hua, Yuan Zhou, Christopher De Sa, Zhiru Zhang, G. Edward Suh

Combining our method with knowledge distillation reduces the compute cost of ResNet-18 by 2. 6$\times$ without accuracy drop on ImageNet.

Knowledge Distillation Network Pruning

Cannot find the paper you are looking for? You can Submit a new open access paper.