no code implementations • EMNLP 2018 • Lifu Huang, Kyunghyun Cho, Boliang Zhang, Heng Ji, Kevin Knight
We construct a multilingual common semantic space based on distributional semantics, where words from multiple languages are projected into a shared space to enable knowledge and resource transfer across languages.
no code implementations • 19 Apr 2018 • Cinjon Resnick, Ilya Kulikov, Kyunghyun Cho, Jason Weston
Interest in emergent communication has recently surged in Machine Learning.
no code implementations • NAACL 2018 • Phu Mon Htut, Samuel R. Bowman, Kyunghyun Cho
In recent years, there have been amazing advances in deep learning methods for machine reading.
no code implementations • ICLR 2018 • Jason Lee, Kyunghyun Cho, Jason Weston, Douwe Kiela
While most machine translation systems to date are trained on large parallel corpora, humans learn language in a different way: by being grounded in an environment and interacting with other humans.
no code implementations • 30 Mar 2018 • Heeyoul Choi, Kyunghyun Cho, Yoshua Bengio
Neural machine translation (NMT) has been a new paradigm in machine translation, and the attention mechanism has become the dominant approach with the state-of-the-art records in many language pairs.
no code implementations • 20 Apr 2017 • Cem M. Deniz, Siyuan Xiang, Spencer Hallyburton, Arakua Welbeck, James S. Babb, Stephen Honig, Kyunghyun Cho, Gregory Chang
However, manual segmentation of MR images of bone is time-consuming, limiting the use of MRI measurements in the clinical practice.
no code implementations • 19 Mar 2018 • Noah Weber, Leena Shekhar, Niranjan Balasubramanian, Kyunghyun Cho
Attention-based neural abstractive summarization systems equipped with copy mechanisms have shown promising results.
no code implementations • 20 May 2017 • Jiatao Gu, Yong Wang, Kyunghyun Cho, Victor O. K. Li
In this paper, we extend an attention-based neural machine translation (NMT) model by allowing it to access an entire training set of parallel sentence pairs even after training.
no code implementations • 26 Feb 2018 • Jake Zhao, Kyunghyun Cho
We propose a retrieval-augmented convolutional network and propose to train it with local mixup, a novel variant of the recently proposed mixup algorithm.
no code implementations • 12 Oct 2017 • Meihao Chen, Zhuoru Lin, Kyunghyun Cho
It is a usual practice to ignore any structural information underlying classes in multi-class classification.
no code implementations • ICLR 2018 • Sean Welleck, Zixin Yao, Yu Gai, Jialin Mao, Zheng Zhang, Kyunghyun Cho
In this paper, we propose a novel multiset loss function by viewing this problem from the perspective of sequential decision making.
no code implementations • NeurIPS 2017 • Sean Welleck, Jialin Mao, Kyunghyun Cho, Zheng Zhang
Humans process visual scenes selectively and sequentially using attention.
no code implementations • 7 Jun 2017 • Keunwoo Choi, George Fazekas, Kyunghyun Cho, Mark Sandler
The results highlight several important aspects of music tagging and neural networks.
no code implementations • 22 Sep 2017 • Tian Wang, Kyunghyun Cho
The goal of personalized history-based recommendation is to automatically output a distribution over all the items given a sequence of previous purchases of a user.
no code implementations • 17 Apr 2017 • Sebastien Jean, Stanislas Lauly, Orhan Firat, Kyunghyun Cho
We propose a neural machine translation architecture that models the surrounding text in addition to the source sentence.
no code implementations • 30 Jun 2016 • Caglar Gulcehre, Sarath Chandar, Kyunghyun Cho, Yoshua Bengio
We investigate the mechanisms and effects of learning to read and write into a memory through experiments on Facebook bAbI tasks using both a feedforward and GRUcontroller.
Ranked #5 on Question Answering on bAbi
no code implementations • 17 Aug 2016 • Keunwoo Choi, George Fazekas, Brian McFee, Kyunghyun Cho, Mark Sandler
Descriptions are often provided along with recommendations to help users' discovery.
no code implementations • 4 Nov 2016 • Hyo-Eun Kim, Sangheum Hwang, Kyunghyun Cho
From the base model, we introduce a semantic noise modeling method which enables class-conditional perturbation on latent space to enhance the representational power of learned latent feature.
no code implementations • 3 Nov 2016 • R. Devon Hjelm, Eswar Damaraju, Kyunghyun Cho, Helmut Laufs, Sergey M. Plis, Vince Calhoun
We introduce a novel recurrent neural network (RNN) approach to account for temporal dynamics and dependencies in brain networks observed via functional magnetic resonance imaging (fMRI).
no code implementations • COLING 2016 • Amrita Saha, Mitesh M. Khapra, Sarath Chandar, Janarthanan Rajendran, Kyunghyun Cho
However, there is no parallel training data available between X and Y but, training data is available between X & Z and Z & Y (as is often the case in many real world applications).
no code implementations • EMNLP 2016 • Orhan Firat, Baskaran Sankaran, Yaser Al-Onaizan, Fatos T. Yarman Vural, Kyunghyun Cho
In this paper, we propose a novel finetuning algorithm for the recently introduced multi-way, mulitlingual neural machine translate that enables zero-resource machine translation.
no code implementations • 8 Jun 2016 • Amjad Almahairi, Kyunghyun Cho, Nizar Habash, Aaron Courville
Neural machine translation has become a major alternative to widely used phrase-based statistical machine translation.
no code implementations • 7 Jun 2016 • Kyunghyun Cho, Masha Esipova
We investigate the potential of attention-based neural machine translation in simultaneous translation.
no code implementations • 12 May 2016 • Kyunghyun Cho
Recent advances in conditional recurrent language modelling have mainly focused on network architectures (e. g., attention mechanism), learning algorithms (e. g., scheduled sampling and sequence-level training) and novel applications (e. g., image/video description generation, speech recognition, etc.)
no code implementations • 19 Nov 2015 • Marcin Moczulski, Kelvin Xu, Aaron Courville, Kyunghyun Cho
Recently there has been growing interest in building active visual object recognizers, as opposed to the usual passive recognizers which classifies a given static image into a predefined set of object categories.
no code implementations • 1 Feb 2016 • Yijun Xiao, Kyunghyun Cho
Document classification tasks were primarily tackled at word level.
no code implementations • NAACL 2016 • Orhan Firat, Kyunghyun Cho, Yoshua Bengio
We propose multi-way, multilingual neural machine translation.
no code implementations • 11 Nov 2015 • Tian Wang, Kyunghyun Cho
In the experi- ments, we evaluate the proposed approach while varying the number of context sentences and observe that the proposed late fusion is superior to the usual way of incorporating additional inputs to the LSTM.
no code implementations • 19 Nov 2015 • Quan Gan, Qipeng Guo, Zheng Zhang, Kyunghyun Cho
In this paper, we propose and study a novel visual object tracking approach based on convolutional networks and recurrent networks.
no code implementations • 4 Jul 2015 • Kyunghyun Cho, Aaron Courville, Yoshua Bengio
Whereas deep neural networks were first mostly used for classification tasks, they are rapidly expanding in the realm of structured output problems, where the observed target is composed of multiple random variables that have a rich joint distribution, given the input.
no code implementations • 9 Feb 2015 • Junyoung Chung, Caglar Gulcehre, Kyunghyun Cho, Yoshua Bengio
In this work, we propose a novel recurrent neural network (RNN) architecture.
no code implementations • 11 Mar 2015 • Caglar Gulcehre, Orhan Firat, Kelvin Xu, Kyunghyun Cho, Loic Barrault, Huei-Chi Lin, Fethi Bougares, Holger Schwenk, Yoshua Bengio
Recent work on end-to-end neural network-based architectures for machine translation has shown promising results for En-Fr and En-De translation.
no code implementations • 19 Dec 2014 • Felix Hill, Kyunghyun Cho, Sebastien Jean, Coline Devin, Yoshua Bengio
Here we investigate the embeddings learned by neural machine translation models, a recently-developed class of neural language model.
no code implementations • 4 Dec 2014 • Jan Chorowski, Dzmitry Bahdanau, Kyunghyun Cho, Yoshua Bengio
We replace the Hidden Markov Model (HMM) which is traditionally used in in continuous speech recognition with a bi-directional recurrent neural network encoder coupled to a recurrent neural network decoder that directly emits a stream of phonemes.
no code implementations • 2 Oct 2014 • Felix Hill, Kyunghyun Cho, Sebastien Jean, Coline Devin, Yoshua Bengio
Neural language models learn word representations that capture rich linguistic and conceptual information.
no code implementations • WS 2014 • Jean Pouget-Abadie, Dzmitry Bahdanau, Bart van Merrienboer, Kyunghyun Cho, Yoshua Bengio
The authors of (Cho et al., 2014a) have shown that the recently introduced neural network translation systems suffer from a significant drop in translation quality when translating long sentences, unlike existing phrase-based translation systems.
no code implementations • 2 Sep 2014 • Li Yao, Sherjil Ozair, Kyunghyun Cho, Yoshua Bengio
Orderless NADEs are trained based on a criterion that stochastically maximizes $P(\mathbf{x})$ with all possible orders of factorizations.
no code implementations • 7 Nov 2013 • Caglar Gulcehre, Kyunghyun Cho, Razvan Pascanu, Yoshua Bengio
In this paper we propose and investigate a novel nonlinear unit, called $L_p$ unit, for deep neural networks.
no code implementations • 17 Jun 2013 • Kyunghyun Cho, Xi Chen
The gesture recognition using motion capture data and depth sensors has recently drawn more attention in vision recognition.
no code implementations • 28 Jun 2014 • Kyunghyun Cho, Yoshua Bengio
Conditional computation has been proposed as a way to increase the capacity of a deep neural network without increasing the amount of computation required, by activating some parameters and computation "on-demand", on a per-example basis.
no code implementations • NeurIPS 2014 • Guido Montúfar, Razvan Pascanu, Kyunghyun Cho, Yoshua Bengio
We study the complexity of functions computable by deep feedforward neural networks with piecewise linear activations in terms of the symmetries and the number of linear regions that they have.
no code implementations • 24 Nov 2013 • Yoshua Bengio, Li Yao, Kyunghyun Cho
Several interesting generative learning algorithms involve a complex probability distribution over many random variables, involving intractable normalization constants or latent variable normalization.
no code implementations • 20 Dec 2013 • Razvan Pascanu, Caglar Gulcehre, Kyunghyun Cho, Yoshua Bengio
Based on this observation, we propose two novel architectures of a deep RNN which are orthogonal to an earlier attempt of stacking multiple recurrent layers to build a deep RNN (Schmidhuber, 1992; El Hihi and Bengio, 1996).
no code implementations • 12 Jun 2013 • Kyunghyun Cho
In this paper, a simple, general method of adding auxiliary stochastic neurons to a multi-layer perceptron is proposed.
no code implementations • 18 Jun 2018 • Amjad Almahairi, Kyle Kastner, Kyunghyun Cho, Aaron Courville
However, interestingly, the greater modeling power offered by the recurrent neural network appears to undermine the model's ability to act as a regularizer of the product representations.
no code implementations • EMNLP 2018 • Jiatao Gu, Yong Wang, Yun Chen, Kyunghyun Cho, Victor O. K. Li
We frame low-resource translation as a meta-learning problem, and we learn to adapt to low-resource languages based on multilingual high-resource language tasks.
no code implementations • ICLR 2019 • Cheolhyoung Lee, Kyunghyun Cho, Wanmo Kang
We empirically verify our result using deep convolutional networks and observe a higher correlation between the gradient stochasticity and the proposed directional uniformity than that against the gradient norm stochasticity, suggesting that the directional statistics of minibatch gradients is a major factor behind SGD.
no code implementations • ACL 2019 • Sean Welleck, Jason Weston, Arthur Szlam, Kyunghyun Cho
Consistency is a long standing issue faced by dialogue models.
no code implementations • EMNLP 2018 • Rujun Han, Michael Gill, Arthur Spirling, Kyunghyun Cho
Conventional word embedding models do not leverage information from document meta-data, and they do not model uncertainty.
no code implementations • WS 2018 • Changhan Wang, Kyunghyun Cho, Douwe Kiela
We describe our work for the CALCS 2018 shared task on named entity recognition on code-switched data.
no code implementations • WS 2017 • Sebastien Jean, Stanislas Lauly, Orhan Firat, Kyunghyun Cho
In this paper we present our systems for the DiscoMT 2017 cross-lingual pronoun prediction shared task.
no code implementations • ICLR 2019 • Cinjon Resnick, Roberta Raileanu, Sanyam Kapoor, Alexander Peysakhovich, Kyunghyun Cho, Joan Bruna
Our contributions are that we analytically characterize the types of environments where Backplay can improve training speed, demonstrate the effectiveness of Backplay both in large grid worlds and a complex four player zero-sum game (Pommerman), and show that Backplay compares favorably to other competitive methods known to improve sample efficiency.
no code implementations • ICLR 2019 • Yu Gai, Zheng Zhang, Kyunghyun Cho
Many important classification performance metrics, e. g. $F$-measure, are non-differentiable and non-decomposable, and are thus unfriendly to gradient descent algorithm.
no code implementations • Dirk Weissenborn, Douwe Kiela, Jason Weston, Kyunghyun Cho
Word inputs tend to be represented as single continuous vectors in deep neural networks.
no code implementations • ICLR 2018 • R. Devon Hjelm, Athul Paul Jacob, Adam Trischler, Gerry Che, Kyunghyun Cho, Yoshua Bengio
We introduce a method for training GANs with discrete data that uses the estimated difference measure from the discriminator to compute importance weights for generated samples, thus providing a policy gradient for training the generator.
no code implementations • ICLR 2018 • Elman Mansimov, Kyunghyun Cho
As this policy does not require any optimization, it allows us to investigate the underlying difficulty of a task without being distracted by optimization difficulty of a learning algorithm.
no code implementations • TACL 2019 • Jiatao Gu, Qi Liu, Kyunghyun Cho
Conventional neural autoregressive decoding commonly assumes a fixed left-to-right generation order, which may be sub-optimal.
no code implementations • 12 Mar 2019 • Sébastien Jean, Kyunghyun Cho
By comparing performance using actual and random contexts, we show that a model trained with the proposed algorithm is more sensitive to the additional context.
no code implementations • 14 May 2019 • Siavash Golkar, Kyunghyun Cho
We introduce a novel algorithm for the detection of possible sample corruption such as mislabeled samples in a training dataset given a small clean validation set.
no code implementations • 24 May 2019 • Iddo Drori, Yamuna Krishnamurthy, Raoni Lourenco, Remi Rampin, Kyunghyun Cho, Claudio Silva, Juliana Freire
Automatic machine learning is an important problem in the forefront of machine learning.
no code implementations • RANLP 2019 • Sean Welleck, Kyunghyun Cho
We propose a method for non-projective dependency parsing by incrementally predicting a set of edges.
no code implementations • 28 May 2019 • Owen Marschall, Kyunghyun Cho, Cristina Savin
To learn useful dynamics on long time scales, neurons must use plasticity rules that account for long-term, circuit-wide effects of synaptic changes.
no code implementations • ACL 2019 • Jiatao Gu, Yong Wang, Kyunghyun Cho, Victor O. K. Li
Zero-shot translation, translating between language pairs on which a Neural Machine Translation (NMT) system has never been trained, is an emergent property when training the system in multilingual settings.
no code implementations • 7 Jun 2019 • Yiqiu Shen, Nan Wu, Jason Phang, Jungkyu Park, Gene Kim, Linda Moy, Kyunghyun Cho, Krzysztof J. Geras
Moreover, both the global structure and local details play important roles in medical image analysis tasks.
no code implementations • 5 Jul 2019 • Owen Marschall, Kyunghyun Cho, Cristina Savin
We present a framework for compactly summarizing many recent results in efficient and/or biologically plausible online training of recurrent neural networks (RNN).
no code implementations • NeurIPS 2019 • Nishant Subramani, Samuel R. Bowman, Kyunghyun Cho
We then investigate the conditions under which a language model can be made to generate a sentence through the identification of a point in such a space and find that it is possible to recover arbitrary sentences nearly perfectly with language models and representations of moderate size without modifying any model parameters.
no code implementations • ACL 2019 • Raphael Shu, Hideki Nakayama, Kyunghyun Cho
In this work, we attempt to obtain diverse translations by using sentence codes to condition the sentence generation.
no code implementations • 30 Jul 2019 • Jungkyu Park, Jason Phang, Yiqiu Shen, Nan Wu, S. Gene Kim, Linda Moy, Kyunghyun Cho, Krzysztof J. Geras
Radiologists typically compare a patient's most recent breast cancer screening exam to their previous ones in making informed diagnoses.
no code implementations • 1 Aug 2019 • Thibault Févry, Jason Phang, Nan Wu, S. Gene Kim, Linda Moy, Kyunghyun Cho, Krzysztof J. Geras
We trained and evaluated a localization-based deep CNN for breast cancer screening exam classification on over 200, 000 exams (over 1, 000, 000 images).
no code implementations • IJCNLP 2019 • Katharina Kann, Kyunghyun Cho, Samuel R. Bowman
Here, we aim to answer the following questions: Does using a development set for early stopping in the low-resource setting influence results as compared to a more realistic alternative, where the number of training epochs is tuned on development languages?
no code implementations • IJCNLP 2019 • Jason Lee, Kyunghyun Cho, Douwe Kiela
Emergent multi-agent communication protocols are very different from natural language and not easily interpretable by humans.
no code implementations • 22 Sep 2019 • Phu Mon Htut, Kyunghyun Cho, Samuel R. Bowman
Latent tree learning(LTL) methods learn to parse sentences using only indirect supervision from a downstream task.
no code implementations • 16 Oct 2019 • Tianxing He, Jun Liu, Kyunghyun Cho, Myle Ott, Bing Liu, James Glass, Fuchun Peng
We find that mix-review effectively regularizes the finetuning process, and the forgetting problem is alleviated to some extent.
no code implementations • WS 2019 • Katharina Kann, Anhad Mohananey, Samuel R. Bowman, Kyunghyun Cho
Recently, neural network models which automatically infer syntactic structure from raw text have started to achieve promising results.
no code implementations • IJCNLP 2019 • Ethan Perez, Siddharth Karamcheti, Rob Fergus, Jason Weston, Douwe Kiela, Kyunghyun Cho
We propose a system that finds the strongest supporting evidence for a given answer to a question, using passage-based question-answering (QA) as a testbed.
no code implementations • 23 Jan 2020 • Rodrigo Nogueira, Zhiying Jiang, Kyunghyun Cho, Jimmy Lin
Citation recommendation systems for the scientific literature, to help authors find papers that should be cited, have the potential to speed up discoveries and uncover new routes for scientific exploration.
no code implementations • ICLR 2020 • Stanislaw Jastrzebski, Maciej Szymczak, Stanislav Fort, Devansh Arpit, Jacek Tabor, Kyunghyun Cho, Krzysztof Geras
We argue for the existence of the "break-even" point on this trajectory, beyond which the curvature of the loss surface and noise in the gradient are implicitly regularized by SGD.
no code implementations • 23 Mar 2020 • Witold Oleszkiewicz, Taro Makino, Stanisław Jastrzębski, Tomasz Trzciński, Linda Moy, Kyunghyun Cho, Laura Heacock, Krzysztof J. Geras
Deep neural networks (DNNs) show promise in breast cancer screening, but their robustness to input perturbations must be better understood before they can be clinically implemented.
no code implementations • 28 Apr 2020 • Katharina Kann, Samuel R. Bowman, Kyunghyun Cho
We propose to cast the task of morphological inflection - mapping a lemma to an indicated inflected form - for resource-poor languages as a meta-learning problem.
no code implementations • WS 2020 • Abhinav Gupta, Cinjon Resnick, Jakob Foerster, Andrew Dai, Kyunghyun Cho
Our hypothesis is that there should be a specific range of model capacity and channel bandwidth that induces compositional structure in the resulting language and consequently encourages systematic generalization.
no code implementations • ICLR 2021 • Shuhei Kurita, Kyunghyun Cho
Vision-and-language navigation (VLN) is a task in which an agent is embodied in a realistic 3D environment and follows an instruction to reach the goal node.
no code implementations • 19 Sep 2020 • Nan Wu, Zhe Huang, Yiqiu Shen, Jungkyu Park, Jason Phang, Taro Makino, S. Gene Kim, Kyunghyun Cho, Laura Heacock, Linda Moy, Krzysztof J. Geras
Breast cancer is the most common cancer in women, and hundreds of thousands of unnecessary biopsies are done around the world at a tremendous cost.
no code implementations • ACL 2020 • Edwin Zhang, Nikhil Gupta, Rodrigo Nogueira, Kyunghyun Cho, Jimmy Lin
The Neural Covidex is a search engine that exploits the latest neural ranking architectures to provide information access to the COVID-19 Open Research Dataset (CORD-19) curated by the Allen Institute for AI.
no code implementations • 11 Nov 2020 • Cinjon Resnick, Or Litany, Hugo Larochelle, Joan Bruna, Kyunghyun Cho
We propose a self-supervised framework to learn scene representations from video that are automatically delineated into objects and background.
no code implementations • EMNLP 2020 • Manling Li, Qi Zeng, Ying Lin, Kyunghyun Cho, Heng Ji, Jonathan May, Nathanael Chambers, Clare Voss
Event schemas can guide our understanding and ability to make predictions with respect to what might happen next.
no code implementations • 3 Dec 2020 • Elham J. Barezi, Iacer Calixto, Kyunghyun Cho, Pascale Fung
These tasks are hard because the label space is usually (i) very large, e. g. thousands or millions of labels, (ii) very sparse, i. e. very few labels apply to each input document, and (iii) highly correlated, meaning that the existence of one label changes the likelihood of predicting all other labels.
no code implementations • 28 Dec 2020 • Stanislaw Jastrzebski, Devansh Arpit, Oliver Astrand, Giancarlo Kerg, Huan Wang, Caiming Xiong, Richard Socher, Kyunghyun Cho, Krzysztof Geras
The early phase of training a deep neural network has a dramatic effect on the local curvature of the loss function.
no code implementations • 23 Jan 2021 • William F. Whitney, Michael Bloesch, Jost Tobias Springenberg, Abbas Abdolmaleki, Kyunghyun Cho, Martin Riedmiller
This causes BBE to be actively detrimental to policy learning in many control tasks.
no code implementations • 1 Feb 2021 • Cinjon Resnick, Or Litany, Cosmas Heiß, Hugo Larochelle, Joan Bruna, Kyunghyun Cho
We propose a self-supervised framework to learn scene representations from video that are automatically delineated into background, characters, and their animations.
no code implementations • ACL 2021 • Clara Vania, Phu Mon Htut, William Huang, Dhara Mungra, Richard Yuanzhe Pang, Jason Phang, Haokun Liu, Kyunghyun Cho, Samuel R. Bowman
Recent years have seen numerous NLP datasets introduced to evaluate the performance of fine-tuned models on natural language understanding tasks.
no code implementations • EACL 2021 • Tianxing He, Jun Liu, Kyunghyun Cho, Myle Ott, Bing Liu, James Glass, Fuchun Peng
We find that mix-review effectively regularizes the finetuning process, and the forgetting problem is alleviated to some extent.
no code implementations • 16 Sep 2021 • Annika Brundyn, Jesse Swanson, Kyunghyun Cho, Doug Kondziolka, Eric Oermann
In the first reader study, a variant of the U-Net that takes as input multiple consecutive video frames and outputs the missing view performs best.
no code implementations • 29 Sep 2021 • Nan Wu, Stanislaw Kamil Jastrzebski, Kyunghyun Cho, Krzysztof J. Geras
We refer to this gain as the conditional utilization rate of the modality.
no code implementations • 29 Sep 2021 • William Alejandro Falcon, Ananya Harsh Jha, Teddy Koker, Kyunghyun Cho
We empirically evaluate the proposed AAVAE on image classification, similar to how recent contrastive and non-contrastive learning algorithms have been evaluated.
no code implementations • 29 Sep 2021 • Cinjon Resnick, Or Litany, Amlan Kar, Karsten Kreis, James Lucas, Kyunghyun Cho, Sanja Fidler
We verify that the prioritized groups found via intervention are challenging for the object detector and show that retraining with data collected from these groups helps inordinately compared to adding more IID data.
no code implementations • WMT (EMNLP) 2021 • Hyojung Han, Seokchan Ahn, Yoonjung Choi, Insoo Chung, Sangha Kim, Kyunghyun Cho
Recent work in simultaneous machine translation is often trained with conventional full sentence translation corpora, leading to either excessive latency or necessity to anticipate as-yet-unarrived words, when dealing with a language pair whose word orders significantly differ.
no code implementations • EMNLP (spnlp) 2020 • Sébastien Jean, Kyunghyun Cho
We seek to maximally use various data sources, such as parallel and monolingual data, to build an effective and efficient document-level translation system.
no code implementations • 3 Nov 2021 • Iddo Drori, Yamuna Krishnamurthy, Remi Rampin, Raoni de Paula Lourenco, Jorge Piazentin Ono, Kyunghyun Cho, Claudio Silva, Juliana Freire
We introduce AlphaD3M, an automatic machine learning (AutoML) system based on meta reinforcement learning using sequence models with self play.
no code implementations • ACL 2022 • Junjie Hu, Hiroaki Hayashi, Kyunghyun Cho, Graham Neubig
It has been shown that machine translation models usually generate poor translations for named entities that are infrequent in the training corpus.
no code implementations • 16 Nov 2021 • Daniel Jiwoong Im, Kyunghyun Cho, Narges Razavian
In this paper, we introduce uniform treatment variational autoencoders (UTVAE) that are trained with uniform treatment distribution using importance sampling and show that using uniform treatment over observational treatment distribution leads to better causal inference by mitigating the distribution shift that occurs from training to test time.
no code implementations • 27 Sep 2018 • Adji B. Dieng, Kyunghyun Cho, David M. Blei, Yann Lecun
Furthermore, the reflective likelihood objective prevents posterior collapse when used to train stochastic auto-encoders with amortized inference.
no code implementations • 27 Sep 2018 • Jason Lee, Kyunghyun Cho, Douwe Kiela
While reinforcement learning (RL) shows a lot of promise for natural language processing—e. g.
no code implementations • MIDL 2019 • Nan Wu, Stanisław Jastrzębski, Jungkyu Park, Linda Moy, Kyunghyun Cho, Krzysztof J. Geras
In breast cancer screening, radiologists make the diagnosis based on images that are taken from two angles.
no code implementations • NeurIPS Workshop Neuro_AI 2019 • Owen Marschall, Kyunghyun Cho, Cristina Savin
To which extent can successful machine learning inform our understanding of biological learning?
no code implementations • 28 Dec 2021 • Yekyung Kim, Seohyeong Jeong, Kyunghyun Cho
Despite the success of mixup in data augmentation, its applicability to natural language processing (NLP) tasks has been limited due to the discrete and variable-length nature of natural languages.
no code implementations • 16 Dec 2021 • Richard Yuanzhe Pang, He He, Kyunghyun Cho
For all three approaches, the generated translations fail to achieve rewards comparable to BSR, but the translation quality approximated by BLEU and BLEURT is similar to the quality of BSR-produced translations.
no code implementations • 8 Feb 2022 • Cinjon Resnick, Or Litany, Amlan Kar, Karsten Kreis, James Lucas, Kyunghyun Cho, Sanja Fidler
Our main contribution is a pseudo-automatic method to discover such groups in foresight by performing causal interventions on simulated scenes.
no code implementations • NAACL 2022 • Seongjin Shin, Sang-Woo Lee, Hwijeen Ahn, Sungdong Kim, HyoungSeok Kim, Boseop Kim, Kyunghyun Cho, Gichang Lee, WooMyoung Park, Jung-Woo Ha, Nako Sung
Many recent studies on large-scale language models have reported successful in-context zero- and few-shot learning ability.
no code implementations • 29 Apr 2022 • Jérémy Scheurer, Jon Ander Campos, Jun Shern Chan, Angelica Chen, Kyunghyun Cho, Ethan Perez
We learn from language feedback on model outputs using a three-step learning algorithm.
no code implementations • 9 May 2022 • Daniel Berenberg, Jae Hyeon Lee, Simon Kelow, Ji Won Park, Andrew Watkins, Vladimir Gligorijević, Richard Bonneau, Stephen Ra, Kyunghyun Cho
We introduce an alternative approach to this guided sampling procedure, multi-segment preserving sampling, that enables the direct inclusion of domain-specific knowledge by designating preserved and non-preserved segments along the input sequence, thereby restricting variation to only select regions.
no code implementations • 20 May 2022 • Juhee Son, Jiho Jin, Haneul Yoo, JinYeong Bak, Kyunghyun Cho, Alice Oh
Built on top of multilingual neural machine translation, H2KE learns to translate a historical document written in Hanja, from both a full dataset of outdated Korean translation and a small dataset of more recently translated contemporary Korean and English.
no code implementations • 5 Jul 2022 • Nathan Ng, Neha Hulkund, Kyunghyun Cho, Marzyeh Ghassemi
Developing and deploying machine learning models safely depends on the ability to characterize and compare their abilities to generalize to new environments.
no code implementations • 28 Aug 2022 • Cal Peyser, Ronny Huang Andrew Rosenberg Tara N. Sainath, Michael Picheny, Kyunghyun Cho
In this paper, we construct a representation learning task based on joint modeling of ASR and TTS, and seek to learn a representation of audio that disentangles that part of the speech signal that is relevant to transcription from that part which is not.
no code implementations • 8 Oct 2022 • Ji Won Park, Samuel Stanton, Saeed Saremi, Andrew Watkins, Henri Dwyer, Vladimir Gligorijevic, Richard Bonneau, Stephen Ra, Kyunghyun Cho
Bayesian optimization offers a sample-efficient framework for navigating the exploration-exploitation trade-off in the vast design space of biological sequences.
no code implementations • 19 Oct 2022 • Nataša Tagasovska, Nathan C. Frey, Andreas Loukas, Isidro Hötzel, Julien Lafrance-Vanasse, Ryan Lewis Kelly, Yan Wu, Arvind Rajpal, Richard Bonneau, Kyunghyun Cho, Stephen Ra, Vladimir Gligorijević
Deep generative models have emerged as a popular machine learning-based approach for inverse design problems in the life sciences.
no code implementations • 20 Dec 2022 • Sang-Woo Lee, Sungdong Kim, Donghyeon Ko, Donghoon Ham, Youngki Hong, Shin Ah Oh, Hyunhoon Jung, Wangkyo Jung, Kyunghyun Cho, Donghyun Kwak, Hyungsuk Noh, WooMyoung Park
Task-oriented dialogue (TOD) systems are mainly based on the slot-filling-based TOD (SF-TOD) framework, in which dialogues are broken down into smaller, controllable units (i. e., slots) to fulfill a specific task.
no code implementations • 11 Jan 2023 • Cal Peyser, Ronny Huang, Tara Sainath, Rohit Prabhavalkar, Michael Picheny, Kyunghyun Cho
Dual learning is a paradigm for semi-supervised machine learning that seeks to leverage unsupervised data by solving two opposite tasks at once.
no code implementations • 8 Feb 2023 • Cheolhyoung Lee, Kyunghyun Cho
We first notice that each parameter configuration in the parameter space corresponds to one particular downstream task of d-way classification.
no code implementations • 19 Apr 2023 • Cal Peyser, Michael Picheny, Kyunghyun Cho, Rohit Prabhavalkar, Ronny Huang, Tara Sainath
Unpaired text and audio injection have emerged as dominant methods for improving ASR performance in the absence of a large labeled corpus.
no code implementations • 23 May 2023 • Angelica Chen, Jason Phang, Alicia Parrish, Vishakh Padmakumar, Chen Zhao, Samuel R. Bowman, Kyunghyun Cho
Large language models (LLMs) have achieved widespread success on a variety of in-context few-shot tasks, but this success is typically evaluated via correctness rather than consistency.
no code implementations • 1 Jun 2023 • Ji Won Park, Nataša Tagasovska, Michael Maser, Stephen Ra, Kyunghyun Cho
At the heart of MOBO is the acquisition function, which determines the next candidate to evaluate by navigating the best compromises among the objectives.
1 code implementation • 23 Jun 2023 • Weizhe Yuan, Kyunghyun Cho, Jason Weston
Natural language (NL) feedback offers rich insights into user experience.
no code implementations • 13 Jul 2023 • Hongyi Zheng, Yixin Zhu, Lavender Yao Jiang, Kyunghyun Cho, Eric Karl Oermann
Recent advances in large language models have led to renewed interest in natural language processing in healthcare using the free text of clinical notes.
no code implementations • 26 Jul 2023 • Richard Yuanzhe Pang, Stephen Roller, Kyunghyun Cho, He He, Jason Weston
We study improving social conversational agents by learning from natural dialogue between users and a deployed model, without extra annotations.
no code implementations • NeurIPS 2023 • Karolis Martinkus, Jan Ludwiczak, Kyunghyun Cho, Wei-Ching Liang, Julien Lafrance-Vanasse, Isidro Hotzel, Arvind Rajpal, Yan Wu, Richard Bonneau, Vladimir Gligorijevic, Andreas Loukas
We introduce AbDiffuser, an equivariant and physics-informed diffusion model for the joint generation of antibody 3D structures and sequences.
no code implementations • 11 Aug 2023 • Cal Peyser, Zhong Meng, Ke Hu, Rohit Prabhavalkar, Andrew Rosenberg, Tara N. Sainath, Michael Picheny, Kyunghyun Cho
The last year has seen astonishing progress in text-prompted image generation premised on the idea of a cross-modal representation space in which the text and image domains are represented jointly.
no code implementations • 18 Aug 2023 • Daniel Jiwoong Im, Kyunghyun Cho
This paper serves as a starting point for machine learning researchers, engineers and students who are interested in but not yet familiar with causal inference.
no code implementations • 4 Sep 2023 • Nathan Ng, Ji Won Park, Jae Hyeon Lee, Ryan Lewis Kelly, Stephen Ra, Kyunghyun Cho
This set embedding represents the "average" of the subreads and can be decoded into a prediction of the clean sequence.
no code implementations • 13 Sep 2023 • Angelica Chen, Ravid Shwartz-Ziv, Kyunghyun Cho, Matthew L. Leavitt, Naomi Saphra
Most interpretability research in NLP focuses on understanding the behavior and features of a fully trained model.
no code implementations • 8 Nov 2023 • Naomi Saphra, Eve Fleisig, Kyunghyun Cho, Adam Lopez
Many NLP researchers are experiencing an existential crisis triggered by the astonishing success of ChatGPT and other systems based on large language models (LLMs).
no code implementations • 7 Dec 2023 • Micah Goldblum, Anima Anandkumar, Richard Baraniuk, Tom Goldstein, Kyunghyun Cho, Zachary C Lipton, Melanie Mitchell, Preetum Nakkiran, Max Welling, Andrew Gordon Wilson
The goal of this series is to chronicle opinions and issues in the field of machine learning as they stand today and as they change over time.
no code implementations • 9 Jan 2024 • Yatong Bai, Utsav Garg, Apaar Shanker, Haoming Zhang, Samyak Parajuli, Erhan Bas, Isidora Filipovic, Amelia N. Chu, Eugenia D Fomitcheva, Elliot Branson, Aerin Kim, Somayeh Sojoudi, Kyunghyun Cho
Vision and vision-language applications of neural networks, such as image classification and captioning, rely on large-scale annotated datasets that require non-trivial data-collecting processes.
no code implementations • 14 Mar 2024 • Sungmin Cha, Kyunghyun Cho
In the Hyperparameter Tuning phase, each algorithm is iteratively trained with different hyperparameter values to find the optimal hyperparameter values.
no code implementations • 2 Apr 2024 • Kang Min Yoo, Jaegeun Han, Sookyo In, Heewon Jeon, Jisu Jeong, Jaewook Kang, Hyunwook Kim, Kyung-Min Kim, Munhyong Kim, Sungju Kim, Donghyun Kwak, Hanock Kwak, Se Jung Kwon, Bado Lee, Dongsoo Lee, Gichang Lee, Jooho Lee, Baeseong Park, Seongjin Shin, Joonsang Yu, Seolki Baek, Sumin Byeon, Eungsup Cho, Dooseok Choe, Jeesung Han, Youngkyun Jin, Hyein Jun, Jaeseung Jung, Chanwoong Kim, jinhong Kim, Jinuk Kim, Dokyeong Lee, Dongwook Park, Jeong Min Sohn, Sujung Han, Jiae Heo, Sungju Hong, Mina Jeon, Hyunhoon Jung, Jungeun Jung, Wangkyo Jung, Chungjoon Kim, Hyeri Kim, Jonghyun Kim, Min Young Kim, Soeun Lee, Joonhee Park, Jieun Shin, Sojin Yang, Jungsoon Yoon, Hwaran Lee, Sanghwan Bae, Jeehwan Cha, Karl Gylleus, Donghoon Ham, Mihak Hong, Youngki Hong, Yunki Hong, Dahyun Jang, Hyojun Jeon, Yujin Jeon, Yeji Jeong, Myunggeun Ji, Yeguk Jin, Chansong Jo, Shinyoung Joo, Seunghwan Jung, Adrian Jungmyung Kim, Byoung Hoon Kim, Hyomin Kim, Jungwhan Kim, Minkyoung Kim, Minseung Kim, Sungdong Kim, Yonghee Kim, Youngjun Kim, Youngkwan Kim, Donghyeon Ko, Dughyun Lee, Ha Young Lee, Jaehong Lee, Jieun Lee, Jonghyun Lee, Jongjin Lee, Min Young Lee, Yehbin Lee, Taehong Min, Yuri Min, Kiyoon Moon, Hyangnam Oh, Jaesun Park, Kyuyon Park, Younghun Park, Hanbae Seo, Seunghyun Seo, Mihyun Sim, Gyubin Son, Matt Yeo, Kyung Hoon Yeom, Wonjoon Yoo, Myungin You, Doheon Ahn, Homin Ahn, Joohee Ahn, Seongmin Ahn, Chanwoo An, Hyeryun An, Junho An, Sang-Min An, Boram Byun, Eunbin Byun, Jongho Cha, Minji Chang, Seunggyu Chang, Haesong Cho, Youngdo Cho, Dalnim Choi, Daseul Choi, Hyoseok Choi, Minseong Choi, Sangho Choi, Seongjae Choi, Wooyong Choi, Sewhan Chun, Dong Young Go, Chiheon Ham, Danbi Han, Jaemin Han, Moonyoung Hong, Sung Bum Hong, Dong-Hyun Hwang, Seongchan Hwang, Jinbae Im, Hyuk Jin Jang, Jaehyung Jang, Jaeni Jang, Sihyeon Jang, Sungwon Jang, Joonha Jeon, Daun Jeong, JoonHyun Jeong, Kyeongseok Jeong, Mini Jeong, Sol Jin, Hanbyeol Jo, Hanju Jo, Minjung Jo, Chaeyoon Jung, Hyungsik Jung, Jaeuk Jung, Ju Hwan Jung, Kwangsun Jung, Seungjae Jung, Soonwon Ka, Donghan Kang, Soyoung Kang, Taeho Kil, Areum Kim, Beomyoung Kim, Byeongwook Kim, Daehee Kim, Dong-Gyun Kim, Donggook Kim, Donghyun Kim, Euna Kim, Eunchul Kim, Geewook Kim, Gyu Ri Kim, Hanbyul Kim, Heesu Kim, Isaac Kim, Jeonghoon Kim, JiHye Kim, Joonghoon Kim, Minjae Kim, Minsub Kim, Pil Hwan Kim, Sammy Kim, Seokhun Kim, Seonghyeon Kim, Soojin Kim, Soong Kim, Soyoon Kim, Sunyoung Kim, TaeHo Kim, Wonho Kim, Yoonsik Kim, You Jin Kim, Yuri Kim, Beomseok Kwon, Ohsung Kwon, Yoo-Hwan Kwon, Anna Lee, Byungwook Lee, Changho Lee, Daun Lee, Dongjae Lee, Ha-Ram Lee, Hodong Lee, Hwiyeong Lee, Hyunmi Lee, Injae Lee, Jaeung Lee, Jeongsang Lee, Jisoo Lee, JongSoo Lee, Joongjae Lee, Juhan Lee, Jung Hyun Lee, Junghoon Lee, Junwoo Lee, Se Yun Lee, Sujin Lee, Sungjae Lee, Sungwoo Lee, Wonjae Lee, Zoo Hyun Lee, Jong Kun Lim, Kun Lim, Taemin Lim, Nuri Na, Jeongyeon Nam, Kyeong-Min Nam, Yeonseog Noh, Biro Oh, Jung-Sik Oh, Solgil Oh, Yeontaek Oh, Boyoun Park, Cheonbok Park, Dongju Park, Hyeonjin Park, Hyun Tae Park, Hyunjung Park, JiHye Park, Jooseok Park, JungHwan Park, Jungsoo Park, Miru Park, Sang Hee Park, Seunghyun Park, Soyoung Park, Taerim Park, Wonkyeong Park, Hyunjoon Ryu, Jeonghun Ryu, Nahyeon Ryu, Soonshin Seo, Suk Min Seo, Yoonjeong Shim, Kyuyong Shin, Wonkwang Shin, Hyun Sim, Woongseob Sim, Hyejin Soh, Bokyong Son, Hyunjun Son, Seulah Son, Chi-Yun Song, Chiyoung Song, Ka Yeon Song, Minchul Song, Seungmin Song, Jisung Wang, Yonggoo Yeo, Myeong Yeon Yi, Moon Bin Yim, Taehwan Yoo, Youngjoon Yoo, Sungmin Yoon, Young Jin Yoon, Hangyeol Yu, Ui Seon Yu, Xingdong Zuo, Jeongin Bae, Joungeun Bae, Hyunsoo Cho, Seonghyun Cho, Yongjin Cho, Taekyoon Choi, Yera Choi, Jiwan Chung, Zhenghui Han, Byeongho Heo, Euisuk Hong, Taebaek Hwang, Seonyeol Im, Sumin Jegal, Sumin Jeon, Yelim Jeong, Yonghyun Jeong, Can Jiang, Juyong Jiang, Jiho Jin, Ara Jo, Younghyun Jo, Hoyoun Jung, Juyoung Jung, Seunghyeong Kang, Dae Hee Kim, Ginam Kim, Hangyeol Kim, Heeseung Kim, Hyojin Kim, Hyojun Kim, Hyun-Ah Kim, Jeehye Kim, Jin-Hwa Kim, Jiseon Kim, Jonghak Kim, Jung Yoon Kim, Rak Yeong Kim, Seongjin Kim, Seoyoon Kim, Sewon Kim, Sooyoung Kim, Sukyoung Kim, Taeyong Kim, Naeun Ko, Bonseung Koo, Heeyoung Kwak, Haena Kwon, Youngjin Kwon, Boram Lee, Bruce W. Lee, Dagyeong Lee, Erin Lee, Euijin Lee, Ha Gyeong Lee, Hyojin Lee, Hyunjeong Lee, Jeeyoon Lee, Jeonghyun Lee, Jongheok Lee, Joonhyung Lee, Junhyuk Lee, Mingu Lee, Nayeon Lee, Sangkyu Lee, Se Young Lee, Seulgi Lee, Seung Jin Lee, Suhyeon Lee, Yeonjae Lee, Yesol Lee, Youngbeom Lee, Yujin Lee, Shaodong Li, Tianyu Liu, Seong-Eun Moon, Taehong Moon, Max-Lasse Nihlenramstroem, Wonseok Oh, Yuri Oh, Hongbeen Park, Hyekyung Park, Jaeho Park, Nohil Park, Sangjin Park, Jiwon Ryu, Miru Ryu, Simo Ryu, Ahreum Seo, Hee Seo, Kangdeok Seo, Jamin Shin, Seungyoun Shin, Heetae Sin, Jiangping Wang, Lei Wang, Ning Xiang, Longxiang Xiao, Jing Xu, Seonyeong Yi, Haanju Yoo, Haneul Yoo, Hwanhee Yoo, Liang Yu, Youngjae Yu, Weijie Yuan, Bo Zeng, Qian Zhou, Kyunghyun Cho, Jung-Woo Ha, Joonsuk Park, Jihyun Hwang, Hyoung Jo Kwon, Soonyong Kwon, Jungyeon Lee, Seungho Lee, Seonghyeon Lim, Hyunkyung Noh, Seungho Choi, Sang-Woo Lee, Jung Hwa Lim, Nako Sung
We introduce HyperCLOVA X, a family of large language models (LLMs) tailored to the Korean language and culture, along with competitive capabilities in English, math, and coding.
no code implementations • 24 Apr 2024 • Saksham Bassi, Duygu Ataman, Kyunghyun Cho
A model's capacity to generalize its knowledge to interpret unseen inputs with different characteristics is crucial to build robust and reliable machine learning systems.
no code implementations • 30 Apr 2024 • Richard Yuanzhe Pang, Weizhe Yuan, Kyunghyun Cho, He He, Sainbayar Sukhbaatar, Jason Weston
Iterative preference optimization methods have recently been shown to perform well for general instruction tuning tasks, but typically make little improvement on reasoning tasks (Yuan et al., 2024, Chen et al., 2024).
1 code implementation • 23 Jun 2023 • Divyam Madaan, Daniel Sodickson, Kyunghyun Cho, Sumit Chopra
However, the image reconstruction process within the MRI pipeline, which requires the use of complex hardware and adjustment of a large number of scanner parameters, is highly susceptible to noise of various forms, resulting in arbitrary artifacts within the images.
1 code implementation • NAACL 2016 • Felix Hill, Kyunghyun Cho, Anna Korhonen
Unsupervised methods for learning distributed representations of words are ubiquitous in today's NLP research, but far less is known about the best ways to learn distributed phrase or sentence representations from unlabelled data.
Ranked #16 on Subjectivity Analysis on SUBJ
1 code implementation • 1 Jun 2019 • Ilia Kulikov, Jason Lee, Kyunghyun Cho
We propose a novel approach for conversation-level inference by explicitly modeling the dialogue partner and running beam search across multiple conversation turns.
1 code implementation • EAMT 2020 • António Góis, Kyunghyun Cho, André Martins
Recent research in neural machine translation has explored flexible generation orders, as an alternative to left-to-right generation.
1 code implementation • COLING 2020 • Jon Ander Campos, Kyunghyun Cho, Arantxa Otegi, Aitor Soroa, Gorka Azkune, Eneko Agirre
The interaction of conversational systems with users poses an exciting opportunity for improving them after deployment, but little evidence has been provided of its feasibility.
1 code implementation • ACL (spnlp) 2021 • Ilia Kulikov, Sean Welleck, Kyunghyun Cho
We propose to study these phenomena by investigating how the modes, or local maxima, of a distribution are maintained throughout the full learning chain of the ground-truth, empirical, learned and decoding-induced distributions, via the newly proposed mode recovery cost.
1 code implementation • 13 Nov 2022 • Grace Yang, Ming Cao, Lavender Y. Jiang, Xujin C. Liu, Alexander T. M. Cheung, Hannah Weiss, David Kurland, Kyunghyun Cho, Eric K. Oermann
We assess the sensitivity score on a set of representative words in the test set using two classifiers trained for hospital readmission classification with similar performance statistics.
1 code implementation • EMNLP 2020 • Sean Welleck, Ilia Kulikov, Jaedeok Kim, Richard Yuanzhe Pang, Kyunghyun Cho
Despite strong performance on a variety of tasks, neural sequence models trained with maximum likelihood have been shown to exhibit issues such as length bias and degenerate repetition.
2 code implementations • 8 Feb 2022 • Taro Makino, Krzysztof J. Geras, Kyunghyun Cho
We propose generative multitask learning (GMTL), a simple and scalable approach to causal representation learning for multitask learning.
1 code implementation • 18 Aug 2023 • Michael Y. Hu, Angelica Chen, Naomi Saphra, Kyunghyun Cho
We use the HMM representation to study phase transitions and identify latent "detour" states that slow down convergence.
1 code implementation • WS 2017 • Kyunghyun Cho
This paper describes a builder entry, named "strawman", to the sentence-level sentiment analysis task of the "Build It, Break It" shared task of the First Workshop on Building Linguistically Generalizable NLP Systems.
1 code implementation • EMNLP 2017 • Jiatao Gu, Kyunghyun Cho, Victor O. K. Li
Instead of trying to build a new decoding algorithm for any specific decoding objective, we propose the idea of trainable decoding algorithm in which we train a decoding algorithm to find a translation that maximizes an arbitrary decoding objective.
1 code implementation • 11 Mar 2019 • Siavash Golkar, Michael Kagan, Kyunghyun Cho
We introduce Continual Learning via Neural Pruning (CLNP), a new method aimed at lifelong learning in fixed capacity models based on neuronal model sparsification.
1 code implementation • 24 Oct 2019 • Cinjon Resnick, Abhinav Gupta, Jakob Foerster, Andrew M. Dai, Kyunghyun Cho
In this paper, we investigate the learning biases that affect the efficacy and compositionality of emergent languages.
1 code implementation • NeurIPS 2016 • R. Devon Hjelm, Kyunghyun Cho, Junyoung Chung, Russ Salakhutdinov, Vince Calhoun, Nebojsa Jojic
Variational methods that rely on a recognition network to approximate the posterior of directed graphical models offer better inference and learning than previous methods.
1 code implementation • 18 Jul 2018 • Cinjon Resnick, Roberta Raileanu, Sanyam Kapoor, Alexander Peysakhovich, Kyunghyun Cho, Joan Bruna
Our contributions are that we analytically characterize the types of environments where Backplay can improve training speed, demonstrate the effectiveness of Backplay both in large grid worlds and a complex four player zero-sum game (Pommerman), and show that Backplay compares favorably to other competitive methods known to improve sample efficiency.
1 code implementation • 16 Dec 2021 • Ilia Kulikov, Maksim Eremeev, Kyunghyun Cho
From these observations, we conclude that the high degree of oversmoothing is the main reason behind the degenerate case of overly probable short sequences in a neural autoregressive model.
1 code implementation • 6 Sep 2021 • Tianxing He, Kyunghyun Cho, James Glass
Prompt-based knowledge probing for 1-hop relations has been used to measure how much world knowledge is stored in pretrained language models.
1 code implementation • 3 Oct 2022 • Eugene Choi, Kyunghyun Cho, Cheolhyoung Lee
We then propose a non-monotonic self-terminating language model, which significantly relaxes the constraint of monotonically increasing termination probability in the originally proposed self-terminating language model by Welleck et al. (2020), to address the issue of non-terminating sequences when using incomplete probable decoding algorithms.
1 code implementation • Asian Chapter of the Association for Computational Linguistics 2020 • Moin Nadeem, Tianxing He, Kyunghyun Cho, James Glass
On the other hand, we find that the set of sampling algorithms that satisfies these properties performs on par with the existing sampling algorithms.
1 code implementation • 15 Feb 2021 • Daniel Jiwoong Im, Cristina Savin, Kyunghyun Cho
Conventional hyperparameter optimization methods are computationally intensive and hard to generalize to scenarios that require dynamically adapting hyperparameters, such as life-long learning.
1 code implementation • 8 Jun 2023 • Nathan C. Frey, Daniel Berenberg, Karina Zadorozhny, Joseph Kleinhenz, Julien Lafrance-Vanasse, Isidro Hotzel, Yan Wu, Stephen Ra, Richard Bonneau, Kyunghyun Cho, Andreas Loukas, Vladimir Gligorijevic, Saeed Saremi
We resolve difficulties in training and sampling from a discrete generative model by learning a smoothed energy function, sampling from the smoothed data manifold with Langevin Markov chain Monte Carlo (MCMC), and projecting back to the true data manifold with one-step denoising.
1 code implementation • WS 2018 • Jasmijn Bastings, Marco Baroni, Jason Weston, Kyunghyun Cho, Douwe Kiela
Lake and Baroni (2018) recently introduced the SCAN data set, which consists of simple commands paired with action sequences and is intended to test the strong generalization abilities of recurrent sequence-to-sequence models.
1 code implementation • 4 Jun 2020 • Sean Welleck, Kyunghyun Cho
Typical approaches to directly optimizing the task loss such as policy gradient and minimum risk training are based around sampling in the sequence space to obtain candidate update directions that are scored based on the loss of a single sequence.
1 code implementation • EMNLP 2020 • Jason Lee, Raphael Shu, Kyunghyun Cho
Given a continuous latent variable model for machine translation (Shu et al., 2020), we train an inference network to approximate the gradient of the marginal log probability of the target sentence, using only the latent variable as input.
1 code implementation • Findings (NAACL) 2022 • Haneul Yoo, Jiho Jin, Juhee Son, JinYeong Bak, Kyunghyun Cho, Alice Oh
Historical records in Korea before the 20th century were primarily written in Hanja, an extinct language based on Chinese characters and not understood by modern Korean or Chinese speakers.
1 code implementation • WS 2019 • Ilia Kulikov, Alexander H. Miller, Kyunghyun Cho, Jason Weston
We investigate the impact of search strategies in neural dialogue modeling.
1 code implementation • 11 May 2023 • Max W. Shen, Emmanuel Bengio, Ehsan Hajiramezanali, Andreas Loukas, Kyunghyun Cho, Tommaso Biancalani
We investigate how to learn better flows, and propose (i) prioritized replay training of high-reward $x$, (ii) relative edge flow policy parametrization, and (iii) a novel guided trajectory balance objective, and show how it can solve a substructure credit assignment problem.
1 code implementation • 16 Nov 2023 • Nicholas Lourie, Kyunghyun Cho, He He
We present the first method to construct valid confidence bands for tuning curves.
1 code implementation • 3 Jul 2016 • Heeyoul Choi, Kyunghyun Cho, Yoshua Bengio
Based on this observation, in this paper we propose to contextualize the word embedding vectors using a nonlinear bag-of-words representation of the source sentence.
1 code implementation • 24 May 2022 • Jeevesh Juneja, Rachit Bansal, Kyunghyun Cho, João Sedoc, Naomi Saphra
It is widely accepted in the mode connectivity literature that when two neural networks are trained similarly on the same data, they are connected by a path through parameter space over which test set accuracy is maintained.
1 code implementation • 20 Nov 2022 • Vlad Sobal, Jyothir S V, Siddhartha Jalagam, Nicolas Carion, Kyunghyun Cho, Yann Lecun
Many common methods for learning a world model for pixel-based environments use generative architectures trained with pixel-level reconstruction objectives.
1 code implementation • 20 Dec 2022 • Tianxing He, Jingyu Zhang, Tianle Wang, Sachin Kumar, Kyunghyun Cho, James Glass, Yulia Tsvetkov
In this work, we explore a useful but often neglected methodology for robustness analysis of text generation evaluation metrics: stress tests with synthetic data.
1 code implementation • ACL 2017 • Akiko Eriguchi, Yoshimasa Tsuruoka, Kyunghyun Cho
There has been relatively little attention to incorporating linguistic prior to neural machine translation.
1 code implementation • 10 Apr 2020 • Edwin Zhang, Nikhil Gupta, Rodrigo Nogueira, Kyunghyun Cho, Jimmy Lin
We present the Neural Covidex, a search engine that exploits the latest neural ranking architectures to provide information access to the COVID-19 Open Research Dataset curated by the Allen Institute for AI.
1 code implementation • IJCNLP 2019 • Laura Graesser, Kyunghyun Cho, Douwe Kiela
In this work, we propose a computational framework in which agents equipped with communication capabilities simultaneously play a series of referential games, where agents are trained using deep reinforcement learning.
1 code implementation • EMNLP 2021 • Manling Li, Sha Li, Zhenhailong Wang, Lifu Huang, Kyunghyun Cho, Heng Ji, Jiawei Han, Clare Voss
We introduce a new concept of Temporal Complex Event Schema: a graph-based schema representation that encompasses events, arguments, temporal connections and argument relations.
1 code implementation • 10 Feb 2022 • Nan Wu, Stanisław Jastrzębski, Kyunghyun Cho, Krzysztof J. Geras
We propose an algorithm to balance the conditional learning speeds between modalities during training and demonstrate that it indeed addresses the issue of greedy learning.
1 code implementation • 28 Mar 2023 • Jérémy Scheurer, Jon Ander Campos, Tomasz Korbak, Jun Shern Chan, Angelica Chen, Kyunghyun Cho, Ethan Perez
Third, finetuning the language model to maximize the likelihood of the chosen refinement given the input.
2 code implementations • EMNLP 2016 • Yasumasa Miyamoto, Kyunghyun Cho
We introduce a recurrent neural network language model (RNN-LM) with long short-term memory (LSTM) units that utilizes both character-level and word-level inputs.
1 code implementation • 5 Jun 2014 • Tapani Raiko, Li Yao, Kyunghyun Cho, Yoshua Bengio
Training of the neural autoregressive density estimator (NADE) can be viewed as doing one step of probabilistic inference on missing values in data.
Ranked #7 on Image Generation on Binarized MNIST
1 code implementation • NeurIPS 2014 • Tapani Raiko, Yao Li, Kyunghyun Cho, Yoshua Bengio
Training of the neural autoregressive density estimator (NADE) can be viewed as doing one step of probabilistic inference on missing values in data.
Ranked #8 on Image Generation on Binarized MNIST
1 code implementation • 29 Apr 2019 • Jihun Oh, Kyunghyun Cho, Joan Bruna
As an efficient and scalable graph neural network, GraphSAGE has enabled an inductive capability for inferring unseen nodes or graphs by aggregating subsampled local neighborhoods and by learning in a mini-batch gradient descent fashion.
1 code implementation • 12 Sep 2019 • Ethan Perez, Siddharth Karamcheti, Rob Fergus, Jason Weston, Douwe Kiela, Kyunghyun Cho
We propose a system that finds the strongest supporting evidence for a given answer to a question, using passage-based question-answering (QA) as a testbed.
4 code implementations • NeurIPS 2014 • Yann Dauphin, Razvan Pascanu, Caglar Gulcehre, Kyunghyun Cho, Surya Ganguli, Yoshua Bengio
Gradient descent or quasi-Newton methods are almost ubiquitously used to perform such minimizations, and it is often thought that a main source of difficulty for these local methods to find the global minimum is the proliferation of local minima with much higher error than the global minimum.
1 code implementation • 29 May 2019 • Elman Mansimov, Alex Wang, Sean Welleck, Kyunghyun Cho
We investigate this problem by proposing a generalized model of sequence generation that unifies decoding in directed and undirected models.
1 code implementation • EACL 2017 • Jiatao Gu, Graham Neubig, Kyunghyun Cho, Victor O. K. Li
Translating in real-time, a. k. a.
1 code implementation • 28 Nov 2020 • Taro Makino, Stanislaw Jastrzebski, Witold Oleszkiewicz, Celin Chacko, Robin Ehrenpreis, Naziya Samreen, Chloe Chhor, Eric Kim, Jiyon Lee, Kristine Pysarenko, Beatriu Reig, Hildegard Toth, Divya Awal, Linda Du, Alice Kim, James Park, Daniel K. Sodickson, Laura Heacock, Linda Moy, Kyunghyun Cho, Krzysztof J. Geras
We compare the two with respect to their robustness to Gaussian low-pass filtering, performing a subgroup analysis on microcalcifications and soft tissue lesions.
1 code implementation • ICLR Workshop Neural_Compression 2021 • Ethan Perez, Douwe Kiela, Kyunghyun Cho
We introduce a method to determine if a certain capability helps to achieve an accurate model of given data.
1 code implementation • 6 Sep 2017 • Keunwoo Choi, György Fazekas, Kyunghyun Cho, Mark Sandler
In this paper, we empirically investigate the effect of audio preprocessing on music tagging with deep neural networks.
2 code implementations • ICLR 2020 • William Whitney, Rajat Agarwal, Kyunghyun Cho, Abhinav Gupta
In this paper we consider self-supervised representation learning to improve sample efficiency in reinforcement learning (RL).
5 code implementations • 2 Feb 2017 • Gilles Louppe, Kyunghyun Cho, Cyril Becot, Kyle Cranmer
Recent progress in applying machine learning for jet physics has been built upon an analogy between calorimeters and images.
1 code implementation • EMNLP (ACL) 2018 • Phu Mon Htut, Kyunghyun Cho, Samuel R. Bowman
A substantial thread of recent work on latent tree learning has attempted to develop neural network models with parse-valued latent variables and train them on non-parsing tasks, in the hope of having them discover interpretable tree structure.
1 code implementation • 26 Jul 2021 • William Falcon, Ananya Harsh Jha, Teddy Koker, Kyunghyun Cho
We empirically evaluate the proposed AASAE on image classification, similar to how recent contrastive and non-contrastive learning algorithms have been evaluated.
1 code implementation • 4 Oct 2023 • Francois Lanusse, Liam Parker, Siavash Golkar, Miles Cranmer, Alberto Bietti, Michael Eickenberg, Geraud Krawezik, Michael McCabe, Ruben Ohana, Mariel Pettee, Bruno Regaldo-Saint Blancard, Tiberiu Tesileanu, Kyunghyun Cho, Shirley Ho
We present AstroCLIP, a strategy to facilitate the construction of astronomical foundation models that bridge the gap between diverse observational modalities.
4 code implementations • 30 Apr 2018 • Seokho Kang, Kyunghyun Cho
Although machine learning has been successfully used to propose novel molecules that satisfy desired properties, it is still challenging to explore a large chemical space efficiently.
1 code implementation • ICLR 2018 • Katrina Evtimova, Andrew Drozdov, Douwe Kiela, Kyunghyun Cho
Inspired by previous work on emergent communication in referential games, we propose a novel multi-modal, multi-step referential game, where the sender and receiver have access to distinct modalities of an object, and their information exchange is bidirectional and of arbitrary duration.