Search Results for author: Zhanda Zhu

Found 4 papers, 2 papers with code

Seesaw: High-throughput LLM Inference via Model Re-sharding

no code implementations9 Mar 2025 Qidong Su, Wei Zhao, Xin Li, Muralidhar Andoorveedu, Chenhao Jiang, Zhanda Zhu, Kevin Song, Christina Giannoula, Gennady Pekhimenko

To improve the efficiency of distributed large language model (LLM) inference, various parallelization strategies, such as tensor and pipeline parallelism, have been proposed.

Computational Efficiency Language Modeling +3

Tempo: Accelerating Transformer-Based Model Training through Memory Footprint Reduction

1 code implementation19 Oct 2022 Muralidhar Andoorveedu, Zhanda Zhu, Bojian Zheng, Gennady Pekhimenko

We implement Tempo and evaluate the throughput, memory usage, and accuracy/loss on the BERT Large pre-training task.

RGB Matters: Learning 7-DoF Grasp Poses on Monocular RGBD Images

1 code implementation3 Mar 2021 Minghao Gou, Hao-Shu Fang, Zhanda Zhu, Sheng Xu, Chenxi Wang, Cewu Lu

In the first stage, an encoder-decoder like convolutional neural network Angle-View Net(AVN) is proposed to predict the SO(3) orientation of the gripper at every location of the image.

Decoder

Cannot find the paper you are looking for? You can Submit a new open access paper.