Search Results for author: Sungjin Ahn

Found 42 papers, 14 papers with code

Learning Attentive Meta-Transfer

no code implementations ICML 2020 Jaesik Yoon, Gautam Singh, Sungjin Ahn

Meta-transfer learning seeks to improve the efficiency of learning a new task via both meta-learning and transfer-learning in a setting with a stream of evolving tasks.

Meta-Learning Transfer Learning

Learning and Simulation in Generative Structured World Models

no code implementations ICML 2020 Zhixuan Lin, Yi-Fu Wu, Skand Peri, Bofeng Fu, Jindong Jiang, Sungjin Ahn

The G-SWM not only unifies the key properties of previous models in a principled framework but also achieves two crucial new abilities, multi-modal uncertainty and situated behavior.

object-detection Object Detection

Parallelized Spatiotemporal Binding

no code implementations26 Feb 2024 Gautam Singh, Yue Wang, Jiawei Yang, Boris Ivanovic, Sungjin Ahn, Marco Pavone, Tong Che

While modern best practices advocate for scalable architectures that support long-range interactions, object-centric models are yet to fully embrace these architectures.

Object

Spatially-Aware Transformer for Embodied Agents

no code implementations23 Feb 2024 Junmo Cho, Jaesik Yoon, Sungjin Ahn

Adopting this approach, we demonstrate that memory utilization efficiency can be improved, leading to enhanced accuracy in various place-centric downstream tasks.

reinforcement-learning

Neural Language of Thought Models

no code implementations2 Feb 2024 Yi-Fu Wu, Minseung Lee, Sungjin Ahn

The Language of Thought Hypothesis suggests that human cognition operates on a structured, language-like system of mental representations.

Image Generation Object +4

Simple Hierarchical Planning with Diffusion

no code implementations5 Jan 2024 Chang Chen, Fei Deng, Kenji Kawaguchi, Caglar Gulcehre, Sungjin Ahn

Diffusion-based generative methods have proven effective in modeling trajectories with offline datasets.

Object-Centric Slot Diffusion

1 code implementation NeurIPS 2023 Jindong Jiang, Fei Deng, Gautam Singh, Sungjin Ahn

The recent success of transformer-based image generative models in object-centric learning highlights the importance of powerful image generators for handling complex scenes.

Image Generation Image Segmentation +2

An Investigation into Pre-Training Object-Centric Representations for Reinforcement Learning

no code implementations9 Feb 2023 Jaesik Yoon, Yi-Fu Wu, Heechul Bae, Sungjin Ahn

In this paper, we investigate the effectiveness of OCR pre-training for image-based reinforcement learning via empirical experiments.

Object Optical Character Recognition (OCR) +5

Neural Systematic Binder

no code implementations2 Nov 2022 Gautam Singh, Yeongbin Kim, Sungjin Ahn

While token-like structured knowledge representations are naturally provided in text, it is elusive how to obtain them for unstructured modalities such as scene images.

Disentanglement Object +2

Simple Unsupervised Object-Centric Learning for Complex and Naturalistic Videos

1 code implementation27 May 2022 Gautam Singh, Yi-Fu Wu, Sungjin Ahn

Unsupervised object-centric learning aims to represent the modular, compositional, and causal structure of a scene as a set of object representations and thereby promises to resolve many critical limitations of traditional single-vector representations such as poor systematic generalization.

Object Systematic Generalization

TransDreamer: Reinforcement Learning with Transformer World Models

no code implementations19 Feb 2022 Chang Chen, Yi-Fu Wu, Jaesik Yoon, Sungjin Ahn

We then share this world model with a transformer-based policy network and obtain stability in training a transformer-based RL agent.

Model-based Reinforcement Learning reinforcement-learning +1

DreamerPro: Reconstruction-Free Model-Based Reinforcement Learning with Prototypical Representations

1 code implementation27 Oct 2021 Fei Deng, Ingook Jang, Sungjin Ahn

Top-performing Model-Based Reinforcement Learning (MBRL) agents, such as Dreamer, learn the world model by reconstructing the image observations.

Model-based Reinforcement Learning reinforcement-learning +1

Illiterate DALL-E Learns to Compose

1 code implementation17 Oct 2021 Gautam Singh, Fei Deng, Sungjin Ahn

In this paper, we propose a simple but novel slot-based autoencoding architecture, called SLATE, for combining the best of both worlds: learning object-centric representations that allows systematic generalization in zero-shot image generation without text.

Image Generation Object +1

Illiterate DALL$\cdot$E Learns to Compose

no code implementations ICLR 2022 Gautam Singh, Fei Deng, Sungjin Ahn

In experiments, we show that this simple architecture achieves zero-shot generation of novel images without text and better quality in generation than the models based on mixture decoders.

Image Generation Systematic Generalization

Generative Video Transformer: Can Objects be the Words?

no code implementations20 Jul 2021 Yi-Fu Wu, Jaesik Yoon, Sungjin Ahn

We compare our model with previous RNN-based approaches as well as other possible video transformer baselines.

Scene Understanding Video Generation

Structured World Belief for Reinforcement Learning in POMDP

no code implementations19 Jul 2021 Gautam Singh, Skand Peri, Junghyun Kim, Hyunseok Kim, Sungjin Ahn

In this paper, we propose Structured World Belief, a model for learning and inference of object-centric belief states.

Inductive Bias Object +3

Generative Scene Graph Networks

no code implementations ICLR 2021 Fei Deng, Zhuo Zhi, Donghun Lee, Sungjin Ahn

We formulate GSGN as a variational autoencoder in which the latent representation is a tree-structured probabilistic scene graph.

Systematic Generalization

Generative Neurosymbolic Machines

1 code implementation NeurIPS 2020 Jindong Jiang, Sungjin Ahn

In this paper, we propose Generative Neurosymbolic Machines, a generative model that combines the benefits of distributed and symbolic representations to support both structured representations of symbolic components and density-based generation.

Image Generation

Improving Generative Imagination in Object-Centric World Models

no code implementations5 Oct 2020 Zhixuan Lin, Yi-Fu Wu, Skand Peri, Bofeng Fu, Jindong Jiang, Sungjin Ahn

Third, a few key abilities for more faithful temporal imagination such as multimodal uncertainty and situation-awareness are missing.

Object object-detection +1

Robustifying Sequential Neural Processes

no code implementations29 Jun 2020 Jaesik Yoon, Gautam Singh, Sungjin Ahn

When tasks change over time, meta-transfer learning seeks to improve the efficiency of learning a new task via both meta-learning and transfer-learning.

Meta-Learning Transfer Learning

ROOTS: Object-Centric Representation and Rendering of 3D Scenes

no code implementations11 Jun 2020 Chang Chen, Fei Deng, Sungjin Ahn

A crucial ability of human intelligence is to build up models of individual 3D objects from partial scene observations.

Object Representation Learning +1

SPACE: Unsupervised Object-Oriented Scene Representation via Spatial Attention and Decomposition

4 code implementations ICLR 2020 Zhixuan Lin, Yi-Fu Wu, Skand Vishwanath Peri, Weihao Sun, Gautam Singh, Fei Deng, Jindong Jiang, Sungjin Ahn

Previous approaches for unsupervised object-oriented scene representation learning are either based on spatial-attention or scene-mixture approaches and limited in scalability which is a main obstacle towards modeling real-world scenes.

Object Representation Learning

Generative Hierarchical Models for Parts, Objects, and Scenes

no code implementations21 Oct 2019 Fei Deng, Zhuo Zhi, Sungjin Ahn

Compositional structures between parts and objects are inherent in natural scenes.

Neural Multisensory Scene Inference

2 code implementations NeurIPS 2019 Jae Hyun Lim, Pedro O. Pinheiro, Negar Rostamzadeh, Christopher Pal, Sungjin Ahn

For embodied agents to infer representations of the underlying 3D physical world they inhabit, they should efficiently combine multisensory cues from numerous trials, e. g., by looking at and touching objects.

Computational Efficiency Representation Learning

SCALOR: Generative World Models with Scalable Object Representations

2 code implementations ICLR 2020 Jindong Jiang, Sepehr Janghorbani, Gerard de Melo, Sungjin Ahn

Scalability in terms of object density in a scene is a primary challenge in unsupervised sequential object-oriented representation learning.

Object Representation Learning

Variational Temporal Abstraction

2 code implementations NeurIPS 2019 Taesup Kim, Sungjin Ahn, Yoshua Bengio

We introduce a variational approach to learning and inference of temporally hierarchical structure and representation for sequential data.

Attentive Sequential Neural Processes

no code implementations25 Sep 2019 Jaesik Yoon, Gautam Singh, Sungjin Ahn

In this paper, we propose the Attentive Sequential Neural Processes (ASNP) that resolve the underfitting in SNP by introducing a novel imaginary context as a latent variable and by applying attention over the imaginary context.

regression

OBJECT-ORIENTED REPRESENTATION OF 3D SCENES

no code implementations25 Sep 2019 Chang Chen, Sungjin Ahn

In this paper, we propose a generative model, called ROOTS (Representation of Object-Oriented Three-dimension Scenes), for unsupervised object-wise 3D-scene decomposition and and rendering.

Disentanglement Object

Sequential Neural Processes

1 code implementation NeurIPS 2019 Gautam Singh, Jaesik Yoon, Youngsung Son, Sungjin Ahn

In this paper, we propose Sequential Neural Processes (SNP) which incorporates a temporal state-transition model of stochastic processes and thus extends its modeling capabilities to dynamic stochastic processes.

Gaussian Processes

Reinforced Imitation Learning from Observations

no code implementations ICLR 2019 Konrad Zolna, Negar Rostamzadeh, Yoshua Bengio, Sungjin Ahn, Pedro O. Pinheiro

Imitation learning is an effective alternative approach to learn a policy when the reward function is sparse.

Imitation Learning

Bayesian Model-Agnostic Meta-Learning

2 code implementations NeurIPS 2018 Taesup Kim, Jaesik Yoon, Ousmane Dia, Sungwoong Kim, Yoshua Bengio, Sungjin Ahn

Learning to infer Bayesian posterior from a few-shot dataset is an important step towards robust meta-learning due to the model uncertainty inherent in the problem.

Active Learning Image Classification +2

Hierarchical Multiscale Recurrent Neural Networks

3 code implementations6 Sep 2016 Junyoung Chung, Sungjin Ahn, Yoshua Bengio

Multiscale recurrent neural networks have been considered as a promising approach to resolve this issue, yet there has been a lack of empirical evidence showing that this type of models can actually capture the temporal dependencies by discovering the latent hierarchical structure of the sequence.

Language Modelling

A Neural Knowledge Language Model

no code implementations1 Aug 2016 Sungjin Ahn, Heeyoul Choi, Tanel Pärnamaa, Yoshua Bengio

Current language models have a significant limitation in the ability to encode and decode factual knowledge.

Language Modelling

Hierarchical Memory Networks

no code implementations24 May 2016 Sarath Chandar, Sungjin Ahn, Hugo Larochelle, Pascal Vincent, Gerald Tesauro, Yoshua Bengio

In this paper, we explore a form of hierarchical memory network, which can be considered as a hybrid between hard and soft attention memory networks.

Hard Attention Question Answering

Pointing the Unknown Words

no code implementations ACL 2016 Caglar Gulcehre, Sungjin Ahn, Ramesh Nallapati, Bo-Wen Zhou, Yoshua Bengio

At each time-step, the decision of which softmax layer to use choose adaptively made by an MLP which is conditioned on the context.~We motivate our work from a psychological evidence that humans naturally have a tendency to point towards objects in the context or the environment when the name of an object is not known.~We observe improvements on two tasks, neural machine translation on the Europarl English to French parallel corpora and text summarization on the Gigaword dataset using our proposed model.

Machine Translation Sentence +2

Denoising Criterion for Variational Auto-Encoding Framework

no code implementations19 Nov 2015 Daniel Jiwoong Im, Sungjin Ahn, Roland Memisevic, Yoshua Bengio

Denoising autoencoders (DAE) are trained to reconstruct their clean inputs with noise injected at the input level, while variational autoencoders (VAE) are trained with noise injected in their stochastic hidden layer, with a regularizer that encourages this noise injection.

Denoising

Scalable MCMC for Mixed Membership Stochastic Blockmodels

no code implementations16 Oct 2015 Wenzhe Li, Sungjin Ahn, Max Welling

We propose a stochastic gradient Markov chain Monte Carlo (SG-MCMC) algorithm for scalable inference in mixed-membership stochastic blockmodels (MMSB).

Variational Inference

Large-Scale Distributed Bayesian Matrix Factorization using Stochastic Gradient MCMC

no code implementations5 Mar 2015 Sungjin Ahn, Anoop Korattikara, Nathan Liu, Suju Rajan, Max Welling

Despite having various attractive qualities such as high prediction accuracy and the ability to quantify uncertainty and avoid over-fitting, Bayesian Matrix Factorization has not been widely adopted because of the prohibitive cost of inference.

Cannot find the paper you are looking for? You can Submit a new open access paper.