1 code implementation • 19 Oct 2023 • Jinheon Baek, Soyeong Jeong, Minki Kang, Jong C. Park, Sung Ju Hwang
Recent Language Models (LMs) have shown impressive capabilities in generating texts with the knowledge internalized in parameters.
no code implementations • 25 Sep 2023 • Minki Kang, Wooseok Han, Eunho Yang
The prosody encoder is specifically designed to model prosodic features that are not captured only with a face image, allowing the face encoder to focus solely on capturing the speaker identity from the face image.
no code implementations • 30 May 2023 • Minki Kang, Jin Myung Kwak, Jinheon Baek, Sung Ju Hwang
To overcome this limitation, we propose SUbgraph Retrieval-augmented GEneration (SURGE), a framework for generating context-relevant and knowledge-grounded dialogues with the KG.
1 code implementation • NeurIPS 2023 • Minki Kang, Seanie Lee, Jinheon Baek, Kenji Kawaguchi, Sung Ju Hwang
Large Language Models (LLMs) have shown promising performance in knowledge-intensive reasoning tasks that require a compound understanding of knowledge.
no code implementations • 23 May 2023 • Minki Kang, Wooseok Han, Sung Ju Hwang, Eunho Yang
Emotional Text-To-Speech (TTS) is an important task in the development of systems (e. g., human-like dialogue agents) that require natural and emotional speech.
no code implementations • 17 Nov 2022 • Minki Kang, Dongchan Min, Sung Ju Hwang
There has been a significant progress in Text-To-Speech (TTS) synthesis technology in recent years, thanks to the advancement in neural generative modeling.
no code implementations • 30 Sep 2022 • Seanie Lee, Minki Kang, Juho Lee, Sung Ju Hwang, Kenji Kawaguchi
Pre-training a large transformer model on a massive amount of unlabeled data and fine-tuning it on labeled datasets for diverse downstream tasks has proven to be a successful strategy, for a variety of vision and natural language processing tasks.
1 code implementation • NAACL 2022 • Minki Kang, Jinheon Baek, Sung Ju Hwang
Pre-trained language models (PLMs) have achieved remarkable success on various natural language understanding tasks.
2 code implementations • NeurIPS 2021 • Jaehyeong Jo, Jinheon Baek, Seul Lee, DongKi Kim, Minki Kang, Sung Ju Hwang
This dual hypergraph construction allows us to apply message-passing techniques for node representations to edges.
1 code implementation • ACL 2021 • Seanie Lee, Minki Kang, Juho Lee, Sung Ju Hwang
QA models based on pretrained language mod-els have achieved remarkable performance on various benchmark datasets. However, QA models do not generalize well to unseen data that falls outside the training distribution, due to distributional shifts. Data augmentation (DA) techniques which drop/replace words have shown to be effective in regularizing the model from overfitting to the training data. Yet, they may adversely affect the QA tasks since they incur semantic changes that may lead to wrong answers for the QA task.
1 code implementation • ICLR 2021 • Jinheon Baek, Minki Kang, Sung Ju Hwang
Graph neural networks have been widely used on modeling graph data, achieving impressive results on node classification and link prediction tasks.
Ranked #1 on Graph Classification on ToxCast
1 code implementation • EMNLP 2020 • Minki Kang, Moonsu Han, Sung Ju Hwang
We propose a method to automatically generate a domain- and task-adaptive maskings of the given text for self-supervised pre-training, such that we can effectively adapt the language model to a particular target task (e. g. question answering).
1 code implementation • ACL 2019 • Moonsu Han, Minki Kang, Hyunwoo Jung, Sung Ju Hwang
We consider a novel question answering (QA) task where the machine needs to read from large streaming data (long documents or videos) without knowing when the questions will be given, which is difficult to solve with existing QA methods due to their lack of scalability.
no code implementations • ICLR 2019 • Hyunwoo Jung, Moonsu Han, Minki Kang, Sungju Hwang
We tackle this problem by proposing a memory network fit for long-term lifelong learning scenario, which we refer to as Long-term Episodic Memory Networks (LEMN), that features a RNN-based retention agent that learns to replace less important memory entries based on the retention probability generated on each entry that is learned to identify data instances of generic importance relative to other memory entries, as well as its historical importance.