Search Results for author: Heegon Jin

Found 2 papers, 1 papers with code

Align-to-Distill: Trainable Attention Alignment for Knowledge Distillation in Neural Machine Translation

1 code implementation3 Mar 2024 Heegon Jin, Seonil Son, Jemin Park, Youngseok Kim, Hyungjong Noh, Yeonsoo Lee

The Attention Alignment Module in A2D performs a dense head-by-head comparison between student and teacher attention heads across layers, turning the combinatorial mapping heuristics into a learning problem.

Knowledge Distillation Machine Translation

Adaptive Attention Link-based Regularization for Vision Transformers

no code implementations25 Nov 2022 Heegon Jin, Jongwon Choi

Although transformer networks are recently employed in various vision tasks with outperforming performance, extensive training data and a lengthy training time are required to train a model to disregard an inductive bias.

Inductive Bias

Cannot find the paper you are looking for? You can Submit a new open access paper.