Search Results for author: Karthikeyan Sankaralingam

Found 3 papers, 1 papers with code

LookupFFN: Making Transformers Compute-lite for CPU inference

1 code implementation12 Mar 2024 Zhanpeng Zeng, Michael Davies, Pranav Pulijala, Karthikeyan Sankaralingam, Vikas Singh

While GPU clusters are the de facto choice for training large deep neural network (DNN) models today, several reasons including ease of workflow, security and cost have led to efforts investigating whether CPUs may be viable for inference in routine use in many sectors of the industry.

Language Modelling

IM-Unpack: Training and Inference with Arbitrarily Low Precision Integers

no code implementations12 Mar 2024 Zhanpeng Zeng, Karthikeyan Sankaralingam, Vikas Singh

A popular strategy is the use of low bit-width integers to approximate the original entries in a matrix.

Open-source Hardware: Opportunities and Challenges

no code implementations7 Jun 2016 Gagan Gupta, Tony Nowatzki, Vinay Gangadhar, Karthikeyan Sankaralingam

Innovation in hardware is slowing due to rising costs of chip design and diminishing benefits from Moore's law and Dennard scaling.

Hardware Architecture Computers and Society

Cannot find the paper you are looking for? You can Submit a new open access paper.