no code implementations • 11 Jul 2024 • Febin Sunny, Amin Shafiee, Abhishek Balasubramaniam, Mahdi Nikdast, Sudeep Pasricha
In this work, we introduce OPIMA, a PIM-based ML accelerator, architected within an optical main memory.
no code implementations • 7 Mar 2024 • Febin Sunny, Ebadollah Taheri, Mahdi Nikdast, Sudeep Pasricha
Modern machine learning (ML) applications are becoming increasingly complex and monolithic (single chip) accelerator architectures cannot keep up with their energy efficiency and throughput demands.
no code implementations • 12 Jan 2024 • Salma Afifi, Febin Sunny, Mahdi Nikdast, Sudeep Pasricha
In the rapidly evolving landscape of artificial intelligence, large language models (LLMs) and graph processing have emerged as transformative technologies for natural language processing (NLP), computer vision, and graph-structured data applications.
no code implementations • 7 Aug 2023 • Amin Shafiee, Sanmitra Banerjee, Krishnendu Chakrabarty, Sudeep Pasricha, Mahdi Nikdast
The proposed models can be applied to any SP-NN architecture with different configurations to analyze the effect of loss and crosstalk.
no code implementations • 4 Jul 2023 • Salma Afifi, Febin Sunny, Amin Shafiee, Mahdi Nikdast, Sudeep Pasricha
Graph neural networks (GNNs) have emerged as a powerful approach for modelling and learning from graph-structured data.
no code implementations • 22 Mar 2023 • Salma Afifi, Febin Sunny, Mahdi Nikdast, Sudeep Pasricha
Transformer neural networks are rapidly being integrated into state-of-the-art solutions for natural language processing (NLP) and computer vision.
no code implementations • 22 Mar 2023 • Febin Sunny, Mahdi Nikdast, Sudeep Pasricha
Emerging AI applications such as ChatGPT, graph convolutional networks, and other deep neural networks require massive computational resources for training and inference.
no code implementations • 28 Jan 2023 • Febin Sunny, Ebadollah Taheri, Mahdi Nikdast, Sudeep Pasricha
Domain-specific machine learning (ML) accelerators such as Google's TPU and Apple's Neural Engine now dominate CPUs and GPUs for energy-efficient ML processing.
no code implementations • 31 Aug 2022 • Febin Sunny, Mahdi Nikdast, Sudeep Pasricha
Recurrent Neural Networks (RNNs) are used in applications that learn dependencies in data sequences, such as speech recognition, human activity recognition, and anomaly detection.
no code implementations • 22 Jul 2022 • Sanmitra Banerjee, Mahdi Nikdast, Krishnendu Chakrabarty
Integrated photonic neural networks (IPNNs) are emerging as promising successors to conventional electronic AI accelerators as they offer substantial improvements in computing speed and energy efficiency.
no code implementations • 17 May 2022 • Febin Sunny, Mahdi Nikdast, Sudeep Pasricha
Parameter quantization in convolutional neural networks (CNNs) can help generate efficient models with lower memory footprint and computational complexity.
no code implementations • 19 Apr 2022 • Asif Mirza, Amin Shafiee, Sanmitra Banerjee, Krishnendu Chakrabarty, Sudeep Pasricha, Mahdi Nikdast
Simulation results for two example SPNNs of different scales under realistic and correlated FPVs indicate that the optimized MZIs can improve the inferencing accuracy by up to 93. 95% for the MNIST handwritten digit dataset -- considered as an example in this paper -- which corresponds to a <0. 5% accuracy loss compared to the variation-free case.
no code implementations • 8 Apr 2022 • Amin Shafiee, Sanmitra Banerjee, Krishnendu Chakrabarty, Sudeep Pasricha, Mahdi Nikdast
Compared to electronic accelerators, integrated silicon-photonic neural networks (SP-NNs) promise higher speed and energy efficiency for emerging artificial-intelligence applications.
no code implementations • 14 Dec 2021 • Sanmitra Banerjee, Mahdi Nikdast, Sudeep Pasricha, Krishnendu Chakrabarty
Singular-value-decomposition-based coherent integrated photonic neural networks (SC-IPNNs) have a large footprint, suffer from high static power consumption for training and inference, and cannot be pruned using conventional DNN pruning techniques.
no code implementations • 11 Dec 2021 • Sanmitra Banerjee, Mahdi Nikdast, Sudeep Pasricha, Krishnendu Chakrabarty
We propose a novel hardware-aware magnitude pruning technique for coherent photonic neural networks.
no code implementations • 9 Sep 2021 • Febin Sunny, Mahdi Nikdast, Sudeep Pasricha
Sparse neural networks can greatly facilitate the deployment of neural networks on resource-constrained platforms as they offer compact model sizes while retaining inference accuracy.
no code implementations • 12 Jul 2021 • Febin P. Sunny, Asif Mirza, Mahdi Nikdast, Sudeep Pasricha
However, mapping sophisticated neural network models on these accelerators still entails significant energy and memory consumption, along with high inference time overhead.
no code implementations • 16 Feb 2021 • Ebadollah Taheri, Ryan G. Kim, Mahdi Nikdast
By lowering the number of vertical connections in fully connected 3D networks-on-chip (NoCs), partially connected 3D NoCs (PC-3DNoCs) help alleviate reliability and fabrication issues.
Distributed, Parallel, and Cluster Computing Hardware Architecture Performance
no code implementations • 13 Feb 2021 • Febin Sunny, Asif Mirza, Mahdi Nikdast, Sudeep Pasricha
Domain-specific neural network accelerators have seen growing interest in recent years due to their improved energy efficiency and inference performance compared to CPUs and GPUs.
no code implementations • 19 Dec 2020 • Sanmitra Banerjee, Mahdi Nikdast, Krishnendu Chakrabarty
Silicon-photonic neural networks (SPNNs) offer substantial improvements in computing speed and energy efficiency compared to their digital electronic counterparts.