Search Results for author: Yingjie Qi

Found 8 papers, 1 papers with code

Graph Neural Networks Automated Design and Deployment on Device-Edge Co-Inference Systems

no code implementations8 Apr 2024 Ao Zhou, Jianlei Yang, Tong Qiao, Yingjie Qi, Zhi Yang, Weisheng Zhao, Chunming Hu

GCoDE abstracts the device communication process into an explicit operation and fuses the search of architecture and the operations mapping in a unified space for joint-optimization.

DDC-PIM: Efficient Algorithm/Architecture Co-design for Doubling Data Capacity of SRAM-based Processing-In-Memory

no code implementations31 Oct 2023 Cenlin Duan, Jianlei Yang, Xiaolin He, Yingjie Qi, Yikun Wang, Yiou Wang, Ziyan He, Bonan Yan, Xueyan Wang, Xiaotao Jia, Weitao Pan, Weisheng Zhao

Processing-in-memory (PIM), as a novel computing paradigm, provides significant performance benefits from the aspect of effective data movement reduction.

Architectural Implications of GNN Aggregation Programming Abstractions

no code implementations18 Oct 2023 Yingjie Qi, Jianlei Yang, Ao Zhou, Tong Qiao, Chunming Hu

Graph neural networks (GNNs) have gained significant popularity due to the powerful capability to extract useful representations from graph data.

Hardware-Aware Graph Neural Network Automated Design for Edge Computing Platforms

no code implementations20 Mar 2023 Ao Zhou, Jianlei Yang, Yingjie Qi, Yumeng Shi, Tong Qiao, Weisheng Zhao, Chunming Hu

Moreover, HGNAS achieves hardware awareness during the GNN architecture design by leveraging a hardware performance predictor, which could balance the GNN model accuracy and efficiency corresponding to the characteristics of targeted devices.

Edge-computing Neural Architecture Search

Eventor: An Efficient Event-Based Monocular Multi-View Stereo Accelerator on FPGA Platform

no code implementations29 Mar 2022 Mingjun Li, Jianlei Yang, Yingjie Qi, Meng Dong, Yuhao Yang, Runze Liu, Weitao Pan, Bei Yu, Weisheng Zhao

In this paper, Eventor is proposed as a fast and efficient EMVS accelerator by realizing the most critical and time-consuming stages including event back-projection and volumetric ray-counting on FPGA.

Quantization

Accelerating CNN Training by Pruning Activation Gradients

no code implementations ECCV 2020 Xucheng Ye, Pengcheng Dai, Junyu Luo, Xin Guo, Yingjie Qi, Jianlei Yang, Yiran Chen

Sparsification is an efficient approach to accelerate CNN inference, but it is challenging to take advantage of sparsity in training procedure because the involved gradients are dynamically changed.

Cannot find the paper you are looking for? You can Submit a new open access paper.