Search Results for author: Hyemin Lee

Found 3 papers, 2 papers with code

A Training-free Sub-quadratic Cost Transformer Model Serving Framework With Hierarchically Pruned Attention

no code implementations14 Jun 2024 Heejun Lee, Geon Park, Youngwan Lee, Jaduk Suh, Jina Kim, Wonyoung Jeong, Bumsik Kim, Hyemin Lee, Myeongjae Jeon, Sung Ju Hwang

In addition to improving the time complexity of the attention mechanism, we further optimize GPU memory usage by implementing KV cache offloading, which stores only $O(\log T)$ tokens on the GPU while maintaining similar decoding throughput.

Question Answering Text Generation

UACANet: Uncertainty Augmented Context Attention for Polyp Segmentation

1 code implementation6 Jul 2021 Taehun Kim, Hyemin Lee, Daijin Kim

We construct a modified version of U-Net shape network with additional encoder and decoder and compute a saliency map in each bottom-up stream prediction module and propagate to the next prediction module.

Decoder Medical Image Segmentation +2

Cannot find the paper you are looking for? You can Submit a new open access paper.