Search Results for author: Yeonsoo Lee

Found 9 papers, 3 papers with code

Align-to-Distill: Trainable Attention Alignment for Knowledge Distillation in Neural Machine Translation

1 code implementation3 Mar 2024 Heegon Jin, Seonil Son, Jemin Park, Youngseok Kim, Hyungjong Noh, Yeonsoo Lee

The Attention Alignment Module in A2D performs a dense head-by-head comparison between student and teacher attention heads across layers, turning the combinatorial mapping heuristics into a learning problem.

Knowledge Distillation Machine Translation

Structured Language Generation Model for Robust Structure Prediction

no code implementations14 Feb 2024 Minho Lee, Junghyun Min, Woochul Lee, Yeonsoo Lee

Previous work in structured prediction (e. g. NER, information extraction) using single model make use of explicit dataset information, which helps boost in-distribution performance but is orthogonal to robust generalization in real-world situations.

NER Structured Prediction +1

Punctuation Restoration Improves Structure Understanding without Supervision

no code implementations13 Feb 2024 Junghyun Min, Minho Lee, Woochul Lee, Yeonsoo Lee

Unsupervised learning objectives like language modeling and de-noising constitute a significant part in producing pre-trained models that perform various downstream applications from natural language understanding to conversational tasks.

Chunking Language Modelling +7

HaRiM$^+$: Evaluating Summary Quality with Hallucination Risk

2 code implementations22 Nov 2022 Seonil Son, Junsoo Park, Jeong-in Hwang, Junghwa Lee, Hyungjong Noh, Yeonsoo Lee

One of the challenges of developing a summarization model arises from the difficulty in measuring the factual inconsistency of the generated text.

Automated Writing Evaluation Hallucination +1

Call for Customized Conversation: Customized Conversation Grounding Persona and Knowledge

2 code implementations16 Dec 2021 Yoonna Jang, Jungwoo Lim, Yuna Hur, Dongsuk Oh, Suhyune Son, Yeonsoo Lee, Donghoon Shin, Seungryong Kim, Heuiseok Lim

Humans usually have conversations by making use of prior knowledge about a topic and background information of the people whom they are talking to.

SANVis: Visual Analytics for Understanding Self-Attention Networks

no code implementations13 Sep 2019 Cheonbok Park, Inyoup Na, Yongjang Jo, Sungbok Shin, Jaehyo Yoo, Bum Chul Kwon, Jian Zhao, Hyungjong Noh, Yeonsoo Lee, Jaegul Choo

Attention networks, a deep neural network architecture inspired by humans' attention mechanism, have seen significant success in image captioning, machine translation, and many other applications.

Image Captioning Machine Translation +2

Question-Aware Sentence Gating Networks for Question and Answering

no code implementations20 Jul 2018 Minjeong Kim, David Keetae Park, Hyungjong Noh, Yeonsoo Lee, Jaegul Choo

Machine comprehension question answering, which finds an answer to the question given a passage, involves high-level reasoning processes of understanding and tracking the relevant contents across various semantic units such as words, phrases, and sentences in a document.

Question Answering Reading Comprehension +1

Cannot find the paper you are looking for? You can Submit a new open access paper.