1 code implementation • 22 May 2024 • Sang Keun Choe, Hwijeen Ahn, Juhan Bae, Kewen Zhao, Minsoo Kang, Youngseog Chung, Adithya Pratapa, Willie Neiswanger, Emma Strubell, Teruko Mitamura, Jeff Schneider, Eduard Hovy, Roger Grosse, Eric Xing
Large language models (LLMs) are trained on a vast amount of human-written data, but data providers often remain uncredited.
1 code implementation • 5 Jul 2022 • Sang Keun Choe, Willie Neiswanger, Pengtao Xie, Eric Xing
Gradient-based multilevel optimization (MLO) has gained attention as a framework for studying numerous problems, ranging from hyperparameter optimization and meta-learning to neural architecture search and reinforcement learning.
2 code implementations • 27 Aug 2020 • Aurick Qiao, Sang Keun Choe, Suhas Jayaram Subramanya, Willie Neiswanger, Qirong Ho, Hao Zhang, Gregory R. Ganger, Eric P. Xing
Some recent schedulers choose job resources for users, but do so without awareness of how DL training can be re-optimized to better utilize the provided resources.
no code implementations • 19 Jun 2020 • Yeojoon Youn, Neil Thistlethwaite, Sang Keun Choe, Jacob Abernethy
We propose a novel approach that resolves many of these issues by relying on a kernel-based non-parametric discriminator that is highly amenable to online training---we call this the Online Kernel-based Generative Adversarial Networks (OKGAN).
no code implementations • WS 2019 • Vikas Raunak, Sang Keun Choe, Quanyang Lu, Yi Xu, Florian Metze
Leveraging the visual modality effectively for Neural Machine Translation (NMT) remains an open problem in computational linguistics.
no code implementations • 1 Dec 2017 • Sungkyun Chang, Juheon Lee, Sang Keun Choe, Kyogu Lee
To do this, we first build the CNN using as an input a cross-similarity matrix generated from a pair of songs.