no code implementations • 18 Feb 2024 • Shirley Anugrah Hayati, Taehee Jung, Tristan Bodding-Long, Sudipta Kar, Abhinav Sethy, Joo-Kyung Kim, Dongyeop Kang
Fine-tuning large language models (LLMs) with a collection of large and diverse instructions has improved the model's generalization to different tasks, even for unseen tasks.
1 code implementation • 17 Feb 2023 • Taehee Jung, Joo-Kyung Kim, Sungjin Lee, Dongyeop Kang
For extreme multi-label classification (XMC), existing classification-based models poorly perform for tail labels and often ignore the semantic relations among labels, like treating "Wikipedia" and "Wiki" as independent and separate labels.
1 code implementation • ACL 2020 • Taehee Jung, Dongyeop Kang, Hua Cheng, Lucas Mentch, Thomas Schaaf
Here we propose an end-to-end training procedure called posterior calibrated (PosCal) training that directly optimizes the objective while minimizing the difference between the predicted and empirical posterior probabilities. We show that PosCal not only helps reduce the calibration error but also improve task performance by penalizing drops in performance of both objectives.
1 code implementation • IJCNLP 2019 • Taehee Jung, Dongyeop Kang, Lucas Mentch, Eduard Hovy
We find that while position exhibits substantial bias in news articles, this is not the case, for example, with academic papers and meeting minutes.