no code implementations • 21 Jan 2019 • Sina Shahhosseini, Ahmad Albaqsami, Masoomeh Jasemi, Nader Bagherzadeh
We evaluated the performance and energy consumption of parallel inference of partitioned models, which showed a 7. 72x speed up of performance and a 2. 73x reduction in the energy used for computing pruned layers of TinyVGG16 in comparison to running the unpruned model on a single accelerator.