no code implementations • ICML 2020 • Jaesik Yoon, Gautam Singh, Sungjin Ahn
Meta-transfer learning seeks to improve the efficiency of learning a new task via both meta-learning and transfer-learning in a setting with a stream of evolving tasks.
no code implementations • ICML 2020 • Zhixuan Lin, Yi-Fu Wu, Skand Peri, Bofeng Fu, Jindong Jiang, Sungjin Ahn
The G-SWM not only unifies the key properties of previous models in a principled framework but also achieves two crucial new abilities, multi-modal uncertainty and situated behavior.
no code implementations • 25 Mar 2025 • Chang Chen, Hany Hamed, Doojin Baek, Taegu Kang, Yoshua Bengio, Sungjin Ahn
This paper tackles a novel problem, extendable long-horizon planning-enabling agents to plan trajectories longer than those in training data without compounding errors.
no code implementations • 11 Feb 2025 • Jaesik Yoon, Hyeonseo Cho, Doojin Baek, Yoshua Bengio, Sungjin Ahn
In this paper, we introduce Monte Carlo Tree Diffusion (MCTD), a novel framework that integrates the generative strength of diffusion models with the adaptive search capabilities of MCTS.
no code implementations • 24 Jan 2025 • Junyeob Baek, Yi-Fu Wu, Gautam Singh, Sungjin Ahn
Furthermore, we show how the modularized concept representations of our model enable compositional imagination, allowing the generation of novel videos by recombining attributes from different objects.
no code implementations • 11 Nov 2024 • Junyeong Park, Junmo Cho, Sungjin Ahn
In this paper, we argue that the primary cause of failure in many low-level controllers is the absence of an episodic memory system.
1 code implementation • 18 Jun 2024 • Jindong Jiang, Fei Deng, Gautam Singh, Minseung Lee, Sungjin Ahn
Unlike conventional SSMs that maintain a monolithic state vector, SlotSSMs maintains the state as a collection of multiple vectors called slots.
1 code implementation • 10 Jun 2024 • Chang Chen, Junyeob Baek, Fei Deng, Kenji Kawaguchi, Caglar Gulcehre, Sungjin Ahn
At the low level, we used a Q-learning based approach called the Q-Performer to accomplish these sub-goals.
1 code implementation • 1 May 2024 • Whie Jung, Jaehoon Yoo, Sungjin Ahn, Seunghoon Hong
Learning compositional representation is a key aspect of object-centric learning as it enables flexible systematic generalization and supports complex visual reasoning.
no code implementations • 29 Feb 2024 • Hany Hamed, Subin Kim, Dongyeong Kim, Jaesik Yoon, Sungjin Ahn
The proposed agent realizes a version of divide-and-conquer-like strategy in dreaming.
no code implementations • 26 Feb 2024 • Gautam Singh, Yue Wang, Jiawei Yang, Boris Ivanovic, Sungjin Ahn, Marco Pavone, Tong Che
While modern best practices advocate for scalable architectures that support long-range interactions, object-centric models are yet to fully embrace these architectures.
1 code implementation • 23 Feb 2024 • Junmo Cho, Jaesik Yoon, Sungjin Ahn
Adopting this approach, we demonstrate that memory utilization efficiency can be improved, leading to enhanced accuracy in various place-centric downstream tasks.
no code implementations • 2 Feb 2024 • Yi-Fu Wu, Minseung Lee, Sungjin Ahn
The Language of Thought Hypothesis suggests that human cognition operates on a structured, language-like system of mental representations.
no code implementations • 5 Jan 2024 • Chang Chen, Fei Deng, Kenji Kawaguchi, Caglar Gulcehre, Sungjin Ahn
Diffusion-based generative methods have proven effective in modeling trajectories with offline datasets.
2 code implementations • NeurIPS 2023 • Jindong Jiang, Fei Deng, Gautam Singh, Sungjin Ahn
The recent success of transformer-based image generative models in object-centric learning highlights the importance of powerful image generators for handling complex scenes.
no code implementations • 9 Feb 2023 • Jaesik Yoon, Yi-Fu Wu, Heechul Bae, Sungjin Ahn
In this paper, we investigate the effectiveness of OCR pre-training for image-based reinforcement learning via empirical experiments.
1 code implementation • 2 Nov 2022 • Gautam Singh, Yeongbin Kim, Sungjin Ahn
While token-like structured knowledge representations are naturally provided in text, it is elusive how to obtain them for unstructured modalities such as scene images.
1 code implementation • 27 May 2022 • Gautam Singh, Yi-Fu Wu, Sungjin Ahn
Unsupervised object-centric learning aims to represent the modular, compositional, and causal structure of a scene as a set of object representations and thereby promises to resolve many critical limitations of traditional single-vector representations such as poor systematic generalization.
no code implementations • 19 Feb 2022 • Chang Chen, Yi-Fu Wu, Jaesik Yoon, Sungjin Ahn
We then share this world model with a transformer-based policy network and obtain stability in training a transformer-based RL agent.
Model-based Reinforcement Learning
reinforcement-learning
+2
1 code implementation • 27 Oct 2021 • Fei Deng, Ingook Jang, Sungjin Ahn
Top-performing Model-Based Reinforcement Learning (MBRL) agents, such as Dreamer, learn the world model by reconstructing the image observations.
Model-based Reinforcement Learning
reinforcement-learning
+2
1 code implementation • 17 Oct 2021 • Gautam Singh, Fei Deng, Sungjin Ahn
In this paper, we propose a simple but novel slot-based autoencoding architecture, called SLATE, for combining the best of both worlds: learning object-centric representations that allows systematic generalization in zero-shot image generation without text.
no code implementations • ICLR 2022 • Gautam Singh, Fei Deng, Sungjin Ahn
In experiments, we show that this simple architecture achieves zero-shot generation of novel images without text and better quality in generation than the models based on mixture decoders.
no code implementations • 20 Jul 2021 • Yi-Fu Wu, Jaesik Yoon, Sungjin Ahn
We compare our model with previous RNN-based approaches as well as other possible video transformer baselines.
no code implementations • 19 Jul 2021 • Gautam Singh, Skand Peri, Junghyun Kim, Hyunseok Kim, Sungjin Ahn
In this paper, we propose Structured World Belief, a model for learning and inference of object-centric belief states.
no code implementations • ICLR 2021 • Fei Deng, Zhuo Zhi, Donghun Lee, Sungjin Ahn
We formulate GSGN as a variational autoencoder in which the latent representation is a tree-structured probabilistic scene graph.
1 code implementation • NeurIPS 2020 • Jindong Jiang, Sungjin Ahn
In this paper, we propose Generative Neurosymbolic Machines, a generative model that combines the benefits of distributed and symbolic representations to support both structured representations of symbolic components and density-based generation.
no code implementations • 5 Oct 2020 • Zhixuan Lin, Yi-Fu Wu, Skand Peri, Bofeng Fu, Jindong Jiang, Sungjin Ahn
Third, a few key abilities for more faithful temporal imagination such as multimodal uncertainty and situation-awareness are missing.
no code implementations • 29 Jun 2020 • Jaesik Yoon, Gautam Singh, Sungjin Ahn
When tasks change over time, meta-transfer learning seeks to improve the efficiency of learning a new task via both meta-learning and transfer-learning.
no code implementations • 11 Jun 2020 • Chang Chen, Fei Deng, Sungjin Ahn
A crucial ability of human intelligence is to build up models of individual 3D objects from partial scene observations.
4 code implementations • ICLR 2020 • Zhixuan Lin, Yi-Fu Wu, Skand Vishwanath Peri, Weihao Sun, Gautam Singh, Fei Deng, Jindong Jiang, Sungjin Ahn
Previous approaches for unsupervised object-oriented scene representation learning are either based on spatial-attention or scene-mixture approaches and limited in scalability which is a main obstacle towards modeling real-world scenes.
no code implementations • 21 Oct 2019 • Fei Deng, Zhuo Zhi, Sungjin Ahn
Compositional structures between parts and objects are inherent in natural scenes.
2 code implementations • ICLR 2020 • Jindong Jiang, Sepehr Janghorbani, Gerard de Melo, Sungjin Ahn
Scalability in terms of object density in a scene is a primary challenge in unsupervised sequential object-oriented representation learning.
2 code implementations • NeurIPS 2019 • Jae Hyun Lim, Pedro O. Pinheiro, Negar Rostamzadeh, Christopher Pal, Sungjin Ahn
For embodied agents to infer representations of the underlying 3D physical world they inhabit, they should efficiently combine multisensory cues from numerous trials, e. g., by looking at and touching objects.
2 code implementations • NeurIPS 2019 • Taesup Kim, Sungjin Ahn, Yoshua Bengio
We introduce a variational approach to learning and inference of temporally hierarchical structure and representation for sequential data.
no code implementations • 25 Sep 2019 • Chang Chen, Sungjin Ahn
In this paper, we propose a generative model, called ROOTS (Representation of Object-Oriented Three-dimension Scenes), for unsupervised object-wise 3D-scene decomposition and and rendering.
no code implementations • 25 Sep 2019 • Jaesik Yoon, Gautam Singh, Sungjin Ahn
In this paper, we propose the Attentive Sequential Neural Processes (ASNP) that resolve the underfitting in SNP by introducing a novel imaginary context as a latent variable and by applying attention over the imaginary context.
1 code implementation • NeurIPS 2019 • Gautam Singh, Jaesik Yoon, Youngsung Son, Sungjin Ahn
In this paper, we propose Sequential Neural Processes (SNP) which incorporates a temporal state-transition model of stochastic processes and thus extends its modeling capabilities to dynamic stochastic processes.
no code implementations • ICLR 2019 • Konrad Zolna, Negar Rostamzadeh, Yoshua Bengio, Sungjin Ahn, Pedro O. Pinheiro
Imitation learning is an effective alternative approach to learn a policy when the reward function is sparse.
no code implementations • 6 Apr 2019 • Konrad Zolna, Negar Rostamzadeh, Yoshua Bengio, Sungjin Ahn, Pedro O. Pinheiro
Imitation learning is an effective alternative approach to learn a policy when the reward function is sparse.
1 code implementation • ICCV 2019 • Pedro O. Pinheiro, Negar Rostamzadeh, Sungjin Ahn
In this paper, we propose a framework to improve over these challenges using adversarial training.
2 code implementations • NeurIPS 2018 • Taesup Kim, Jaesik Yoon, Ousmane Dia, Sungwoong Kim, Yoshua Bengio, Sungjin Ahn
Learning to infer Bayesian posterior from a few-shot dataset is an important step towards robust meta-learning due to the model uncertainty inherent in the problem.
3 code implementations • 6 Sep 2016 • Junyoung Chung, Sungjin Ahn, Yoshua Bengio
Multiscale recurrent neural networks have been considered as a promising approach to resolve this issue, yet there has been a lack of empirical evidence showing that this type of models can actually capture the temporal dependencies by discovering the latent hierarchical structure of the sequence.
Ranked #19 on
Language Modelling
on Text8
no code implementations • 1 Aug 2016 • Sungjin Ahn, Heeyoul Choi, Tanel Pärnamaa, Yoshua Bengio
Current language models have a significant limitation in the ability to encode and decode factual knowledge.
no code implementations • 24 May 2016 • Sarath Chandar, Sungjin Ahn, Hugo Larochelle, Pascal Vincent, Gerald Tesauro, Yoshua Bengio
In this paper, we explore a form of hierarchical memory network, which can be considered as a hybrid between hard and soft attention memory networks.
no code implementations • ACL 2016 • Caglar Gulcehre, Sungjin Ahn, Ramesh Nallapati, Bo-Wen Zhou, Yoshua Bengio
At each time-step, the decision of which softmax layer to use choose adaptively made by an MLP which is conditioned on the context.~We motivate our work from a psychological evidence that humans naturally have a tendency to point towards objects in the context or the environment when the name of an object is not known.~We observe improvements on two tasks, neural machine translation on the Europarl English to French parallel corpora and text summarization on the Gigaword dataset using our proposed model.
1 code implementation • ACL 2016 • Iulian Vlad Serban, Alberto García-Durán, Caglar Gulcehre, Sungjin Ahn, Sarath Chandar, Aaron Courville, Yoshua Bengio
Over the past decade, large-scale supervised learning corpora have enabled machine learning researchers to make substantial advances.
no code implementations • 19 Nov 2015 • Daniel Jiwoong Im, Sungjin Ahn, Roland Memisevic, Yoshua Bengio
Denoising autoencoders (DAE) are trained to reconstruct their clean inputs with noise injected at the input level, while variational autoencoders (VAE) are trained with noise injected in their stochastic hidden layer, with a regularizer that encourages this noise injection.
no code implementations • 16 Oct 2015 • Wenzhe Li, Sungjin Ahn, Max Welling
We propose a stochastic gradient Markov chain Monte Carlo (SG-MCMC) algorithm for scalable inference in mixed-membership stochastic blockmodels (MMSB).
no code implementations • 5 Mar 2015 • Sungjin Ahn, Anoop Korattikara, Nathan Liu, Suju Rajan, Max Welling
Despite having various attractive qualities such as high prediction accuracy and the ability to quantify uncertainty and avoid over-fitting, Bayesian Matrix Factorization has not been widely adopted because of the prohibitive cost of inference.