Search Results for author: Yusuke Nagasaka

Found 2 papers, 2 papers with code

Batched Sparse Matrix Multiplication for Accelerating Graph Convolutional Networks

1 code implementation27 Mar 2019 Yusuke Nagasaka, Akira Nukada, Ryosuke Kojima, Satoshi Matsuoka

We evaluated the performance of the GCNs application on TSUBAME3. 0 implementing NVIDIA Tesla P100 GPU, and our batched approach shows significant speedups of up to 1. 59x and 1. 37x in training and inference, respectively.

Distributed, Parallel, and Cluster Computing

High-performance sparse matrix-matrix products on Intel KNL and multicore architectures

1 code implementation5 Apr 2018 Yusuke Nagasaka, Satoshi Matsuoka, Ariful Azad, Aydın Buluç

Our hash-table and heap-based algorithms are showing significant speedups from libraries in the majority of the cases while different algorithms dominate the other scenarios with different matrix size, sparsity, compression factor and operation type.

Distributed, Parallel, and Cluster Computing

Cannot find the paper you are looking for? You can Submit a new open access paper.