Search Results for author: Jinkyoo Park

Found 60 papers, 20 papers with code

Ensembling Prioritized Hybrid Policies for Multi-agent Pathfinding

1 code implementation12 Mar 2024 Huijie Tang, Federico Berto, Jinkyoo Park

We first propose a selective communication block to gather richer information for better agent coordination within multi-agent environments and train the model with a Q-learning-based algorithm.

Multi-Agent Path Finding Multi-agent Reinforcement Learning +1

Ant Colony Sampling with GFlowNets for Combinatorial Optimization

2 code implementations11 Mar 2024 Minsu Kim, Sanghyeok Choi, Jiwoo Son, Hyeonah Kim, Jinkyoo Park, Yoshua Bengio

This paper introduces the Generative Flow Ant Colony Sampler (GFACS), a novel neural-guided meta-heuristic algorithm for combinatorial optimization.

Combinatorial Optimization

ELA: Exploited Level Augmentation for Offline Learning in Zero-Sum Games

no code implementations28 Feb 2024 Shiqi Lei, Kanghoon Lee, Linjing Li, Jinkyoo Park, Jiachen Li

Offline learning has become widely used due to its ability to derive effective policies from offline datasets gathered by expert demonstrators without interacting with the environment directly.

Imitation Learning

HiMAP: Learning Heuristics-Informed Policies for Large-Scale Multi-Agent Pathfinding

1 code implementation23 Feb 2024 Huijie Tang, Federico Berto, Zihan Ma, Chuanbo Hua, Kyuree Ahn, Jinkyoo Park

With a simple training scheme and implementation, HiMAP demonstrates competitive results in terms of success rate and scalability in the field of imitation-learning-only MAPF, showing the potential of imitation-learning-only MAPF equipped with inference techniques.

Imitation Learning Reinforcement Learning (RL)

Genetic-guided GFlowNets: Advancing in Practical Molecular Optimization Benchmark

no code implementations5 Feb 2024 Hyeonah Kim, Minsu Kim, Sanghyeok Choi, Jinkyoo Park

This paper proposes a novel variant of GFlowNet, genetic-guided GFlowNet (Genetic GFN), which integrates an iterative genetic search into GFlowNet.

Bayesian Optimization

Multi-Agent Dynamic Relational Reasoning for Social Robot Navigation

no code implementations22 Jan 2024 Jiachen Li, Chuanbo Hua, Hengbo Ma, Jinkyoo Park, Victoria Dax, Mykel J. Kochenderfer

In this paper, we propose a systematic relational reasoning approach with explicit inference of the underlying dynamically evolving relational structures, and we demonstrate its effectiveness for multi-agent trajectory prediction and social robot navigation.

Relational Reasoning Robot Navigation +2

Interactive Autonomous Navigation with Internal State Inference and Interactivity Estimation

no code implementations27 Nov 2023 Jiachen Li, David Isele, Kanghoon Lee, Jinkyoo Park, Kikuo Fujimura, Mykel J. Kochenderfer

Moreover, we propose an interactivity estimation mechanism based on the difference between predicted trajectories in these two situations, which indicates the degree of influence of the ego agent on other agents.

Autonomous Navigation counterfactual +4

Genetic Algorithms with Neural Cost Predictor for Solving Hierarchical Vehicle Routing Problems

no code implementations22 Oct 2023 Abhay Sobhanan, Junyoung Park, Jinkyoo Park, Changhyun Kwon

For each higher-level decision candidate, we predict the objective function values of the underlying vehicle routing problems using a pre-trained graph neural network without actually solving the routing problems.

Learning to Scale Logits for Temperature-Conditional GFlowNets

1 code implementation4 Oct 2023 Minsu Kim, Joohwan Ko, Taeyoung Yun, Dinghuai Zhang, Ling Pan, Woochang Kim, Jinkyoo Park, Emmanuel Bengio, Yoshua Bengio

We find that the challenge is greatly reduced if a learned function of the temperature is used to scale the policy's logits directly.

Local Search GFlowNets

2 code implementations4 Oct 2023 Minsu Kim, Taeyoung Yun, Emmanuel Bengio, Dinghuai Zhang, Yoshua Bengio, Sungsoo Ahn, Jinkyoo Park

Generative Flow Networks (GFlowNets) are amortized sampling methods that learn a distribution over discrete objects proportional to their rewards.

Robust Driving Policy Learning with Guided Meta Reinforcement Learning

no code implementations19 Jul 2023 Kanghoon Lee, Jiachen Li, David Isele, Jinkyoo Park, Kikuo Fujimura, Mykel J. Kochenderfer

Although deep reinforcement learning (DRL) has shown promising results for autonomous navigation in interactive traffic scenarios, existing work typically adopts a fixed behavior policy to control social vehicles in the training environment.

Autonomous Navigation Meta Reinforcement Learning +1

A Neural Separation Algorithm for the Rounded Capacity Inequalities

1 code implementation29 Jun 2023 Hyeonah Kim, Jinkyoo Park, Changhyun Kwon

We design a learning-based separation heuristic algorithm with graph coarsening that learns the solutions of the exact separation problem with a graph neural network (GNN), which is trained with small instances of 50 to 100 customers.

Bootstrapped Training of Score-Conditioned Generator for Offline Design of Biological Sequences

1 code implementation NeurIPS 2023 Minsu Kim, Federico Berto, Sungsoo Ahn, Jinkyoo Park

The subsequent stage involves bootstrapping, which augments the training dataset with self-generated data labeled by a proxy score function.

Equity-Transformer: Solving NP-hard Min-Max Routing Problems as Sequential Generation with Equity Context

1 code implementation5 Jun 2023 Jiwoo Son, Minsu Kim, Sanghyeok Choi, Hyeonah Kim, Jinkyoo Park

Notably, our method achieves significant reductions of runtime, approximately 335 times, and cost values of about 53\% compared to a competitive heuristic (LKH3) in the case of 100 vehicles with 1, 000 cities of mTSP.

Decision Making Traveling Salesman Problem

Generating Dispatching Rules for the Interrupting Swap-Allowed Blocking Job Shop Problem Using Graph Neural Network and Reinforcement Learning

no code implementations5 Feb 2023 Vivian W. H. Wong, Sang Hun Kim, Junyoung Park, Jinkyoo Park, Kincho H. Law

The interrupting swap-allowed blocking job shop problem (ISBJSSP) is a complex scheduling problem that is able to model many manufacturing planning and logistics applications realistically by addressing both the lack of storage capacity and unforeseen production interruptions.

Blocking Scheduling

Learning context-aware adaptive solvers to accelerate quadratic programming

no code implementations22 Nov 2022 Haewon Jung, Junyoung Park, Jinkyoo Park

Convex quadratic programming (QP) is an important sub-field of mathematical optimization.

Bayesian Convolutional Deep Sets with Task-Dependent Stationary Prior

no code implementations22 Oct 2022 Yohan Jung, Jinkyoo Park

To remedy this issue, we introduce Bayesian convolutional deep sets that construct the random translation equivariant functional representations with stationary prior.

Inductive Bias Time Series +2

EvolveHypergraph: Group-Aware Dynamic Relational Reasoning for Trajectory Prediction

no code implementations10 Aug 2022 Jiachen Li, Chuanbo Hua, Jinkyoo Park, Hengbo Ma, Victoria Dax, Mykel J. Kochenderfer

While the modeling of pair-wise relations has been widely studied in multi-agent interacting systems, its ability to capture higher-level and larger-scale group-wise activities is limited.

Relation Relational Reasoning +1

Neuro CROSS exchange: Learning to CROSS exchange to solve realistic vehicle routing problems

no code implementations6 Jun 2022 Minjun Kim, Junyoung Park, Jinkyoo Park

Inspired by CE, we propose Neuro CE (NCE), a fundamental operator of learned meta-heuristic, to solve various VRPs while overcoming the limitations of CE (i. e., the expensive $\mathcal{O}(n^4)$ search cost).

Meta-SysId: A Meta-Learning Approach for Simultaneous Identification and Prediction

no code implementations1 Jun 2022 Junyoung Park, Federico Berto, Arec Jamgochian, Mykel J. Kochenderfer, Jinkyoo Park

In this paper, we propose Meta-SysId, a meta-learning approach to model sets of systems that have behavior governed by common but unknown laws and that differentiate themselves by their context.

Meta-Learning regression +3

DevFormer: A Symmetric Transformer for Context-Aware Device Placement

2 code implementations26 May 2022 Haeyeon Kim, Minsu Kim, Federico Berto, Joungho Kim, Jinkyoo Park

In this paper, we present DevFormer, a novel transformer-based architecture for addressing the complex and computationally demanding problem of hardware design optimization.

Combinatorial Optimization Meta-Learning

Sym-NCO: Leveraging Symmetricity for Neural Combinatorial Optimization

1 code implementation26 May 2022 Minsu Kim, Junyoung Park, Jinkyoo Park

Deep reinforcement learning (DRL)-based combinatorial optimization (CO) methods (i. e., DRL-NCO) have shown significant merit over the conventional CO solvers as DRL-NCO is capable of learning CO solvers less relying on problem-specific expert domain knowledge (heuristic method) and supervised labeled data (supervised learning method).

Combinatorial Optimization Traveling Salesman Problem

Neural Solvers for Fast and Accurate Numerical Optimal Control

1 code implementation NeurIPS Workshop DLDE 2021 Federico Berto, Stefano Massaroli, Michael Poli, Jinkyoo Park

Synthesizing optimal controllers for dynamical systems often involves solving optimization problems with hard real-time constraints.

Learning Collaborative Policies to Solve NP-hard Routing Problems

1 code implementation NeurIPS 2021 Minsu Kim, Jinkyoo Park, Joungho Kim

Recently, deep reinforcement learning (DRL) frameworks have shown potential for solving NP-hard routing problems such as the traveling salesman problem (TSP) without problem-specific expert knowledge.

Traveling Salesman Problem

Input Convex Graph Neural Networks: An Application to Optimal Control and Design Optimization

no code implementations29 Sep 2021 Junyoung Park, Chihyeon Song, Jinkyoo Park

On the physical heat diffusion, we further apply ICGNN to solve a design optimization problem, which seeks to find the optimal heater allocations while considering the optimal operation of the heaters, by using a gradient-based method.

Decision Making

A molecular hypergraph convolutional network with functional group information

no code implementations29 Sep 2021 Junyoung Park, Fangying Chen, Jinkyoo Park

We show that our model is able to outperform other baseline methods for most of the datasets.

Property Prediction

LPMARL: Linear Programming based Implicit Task Assigment for Hiearchical Multi-Agent Reinforcement Learning

no code implementations29 Sep 2021 Kyuree Ahn, Jinkyoo Park

Earlier studies have tried to resolve this issue by using hierarchical MARL to decompose the main task into subproblems or employing an intrinsic reward to induce interactions for learning an effective policy.

Multi-agent Reinforcement Learning reinforcement-learning +1

Neural Hybrid Automata: Learning Dynamics with Multiple Modes and Stochastic Transitions

no code implementations NeurIPS 2021 Michael Poli, Stefano Massaroli, Luca Scimeca, Seong Joon Oh, Sanghyuk Chun, Atsushi Yamashita, Hajime Asama, Jinkyoo Park, Animesh Garg

Effective control and prediction of dynamical systems often require appropriate handling of continuous-time and discrete, event-triggered processes.

Learning Stochastic Optimal Policies via Gradient Descent

no code implementations7 Jun 2021 Stefano Massaroli, Michael Poli, Stefano Peluchetti, Jinkyoo Park, Atsushi Yamashita, Hajime Asama

We systematically develop a learning-based treatment of stochastic optimal control (SOC), relying on direct optimization of parametric control policies.

Portfolio Optimization

ScheduleNet: Learn to solve multi-agent scheduling problems with reinforcement learning

no code implementations6 Jun 2021 Junyoung Park, Sanjar Bakhtiyar, Jinkyoo Park

We propose ScheduleNet, a RL-based real-time scheduler, that can solve various types of multi-agent scheduling problems.

Decision Making Graph Attention +4

Convergent Graph Solvers

1 code implementation ICLR 2022 Junyoung Park, Jinhyun Choo, Jinkyoo Park

We propose the convergent graph solver (CGS), a deep learning method that learns iterative mappings to predict the properties of a graph system at its stationary state (fixed point) with guaranteed convergence.

Graph Classification

Learning to schedule job-shop problems: Representation and policy learning using graph neural network and reinforcement learning

no code implementations2 Jun 2021 Junyoung Park, Jaehyeong Chun, Sang Hun Kim, Youngkook Kim, Jinkyoo Park

In solving the formulated problem, the proposed framework employs a GNN to learn that node features that embed the spatial structure of the JSSP represented as a graph (representation learning) and derive the optimum scheduling policy that maps the embedded node features to the best scheduling action (policy learning).

Decision Making Graph Representation Learning +2

Cooperative and Competitive Biases for Multi-Agent Reinforcement Learning

no code implementations18 Jan 2021 Heechang Ryu, Hayong Shin, Jinkyoo Park

We propose an algorithm that boosts MARL training using the biased action information of other agents based on a friend-or-foe concept.

Multi-agent Reinforcement Learning reinforcement-learning +1

Optimal Energy Shaping via Neural Approximators

no code implementations14 Jan 2021 Stefano Massaroli, Michael Poli, Federico Califano, Jinkyoo Park, Atsushi Yamashita, Hajime Asama

We introduce optimal energy shaping as an enhancement of classical passivity-based control methods.

Embedding a random graph via GNN: mean-field inference theory and RL applications to NP-Hard multi-robot/machine scheduling

no code implementations1 Jan 2021 Hyunwook Kang, SEUNGWOO SCHIN, James Morrison, Jinkyoo Park

Experimental results on solving NP-hard MRRC problems (and IMPS in the Appendix) highlight the near-optimality and transferability of the proposed methods.

Scheduling

ScheduleNet: Learn to Solve MinMax mTSP Using Reinforcement Learning with Delayed Reward

no code implementations1 Jan 2021 Junyoung Park, Sanzhar Bakhtiyarov, Jinkyoo Park

From the RL perspective, Minmax mTSP raises several significant challenges, such as the cooperation of multiple workers and the need for a well-engineered reward function.

Combinatorial Optimization reinforcement-learning +1

Scalable Hybrid Hidden Markov Model with Gaussian Process Emission for Sequential Time-series Observations

no code implementations pproximateinference AABI Symposium 2021 Yohan Jung, Jinkyoo Park

We then propose a scalable learning method to train the HMM-GPSM model using large-scale data having (1) long sequences of state transitions and (2) a large number of time-series observations for each hidden state.

Time Series Time Series Analysis +1

TorchDyn: A Neural Differential Equations Library

no code implementations20 Sep 2020 Michael Poli, Stefano Massaroli, Atsushi Yamashita, Hajime Asama, Jinkyoo Park

Continuous-depth learning has recently emerged as a novel perspective on deep learning, improving performance in tasks related to dynamical systems and density estimation.

Density Estimation

REMAX: Relational Representation for Multi-Agent Exploration

no code implementations12 Aug 2020 Heechang Ryu, Hayong Shin, Jinkyoo Park

To train the MARL model effectively without designing the intrinsic reward, we propose a learning-based exploration strategy to generate the initial states of a game.

Multi-agent Reinforcement Learning

Hypersolvers: Toward Fast Continuous-Depth Models

1 code implementation NeurIPS 2020 Michael Poli, Stefano Massaroli, Atsushi Yamashita, Hajime Asama, Jinkyoo Park

The infinite-depth paradigm pioneered by Neural ODEs has launched a renaissance in the search for novel dynamical system-inspired deep learning primitives; however, their utilization in problems of non-trivial size has often proved impossible due to poor computational scalability.

Approximate Inference for Spectral Mixture Kernel

no code implementations12 Jun 2020 Yohan Jung, Kyungwoo Song, Jinkyoo Park

To improve the training, we propose an approximate Bayesian inference for the SM kernel.

Bayesian Inference Variational Inference

Stable Neural Flows

no code implementations18 Mar 2020 Stefano Massaroli, Michael Poli, Michelangelo Bin, Jinkyoo Park, Atsushi Yamashita, Hajime Asama

We introduce a provably stable variant of neural ordinary differential equations (neural ODEs) whose trajectories evolve on an energy functional parametrised by a neural network.

Port-Hamiltonian Gradient Flows

no code implementations ICLR Workshop DeepDiffEq 2019 Michael Poli, Stefano Massaroli, Atsushi Yamashita, Hajime Asama, Jinkyoo Park

In this paper we present a general framework for continuous--time gradient descent, often referred to as gradient flow.

Dissecting Neural ODEs

1 code implementation NeurIPS 2020 Stefano Massaroli, Michael Poli, Jinkyoo Park, Atsushi Yamashita, Hajime Asama

Continuous deep learning architectures have recently re-emerged as Neural Ordinary Differential Equations (Neural ODEs).

Graph Neural Ordinary Differential Equations

1 code implementation18 Nov 2019 Michael Poli, Stefano Massaroli, Junyoung Park, Atsushi Yamashita, Hajime Asama, Jinkyoo Park

We introduce the framework of continuous--depth graph neural networks (GNNs).

Spectral Mixture Kernel Approximation Using Reparameterized Random Fourier Feature

no code implementations pproximateinference AABI Symposium 2019 Yohan Jung, Jinkyoo Park

We propose a method for Spectral Mixture kernel approximation using the Reparameterized Random Fourier Feature (R-RFF) in the sense of both general parameter and natural parameter view.

Multi-Agent Actor-Critic with Hierarchical Graph Attention Network

no code implementations27 Sep 2019 Heechang Ryu, Hayong Shin, Jinkyoo Park

Most previous studies on multi-agent reinforcement learning focus on deriving decentralized and cooperative policies to maximize a common reward and rarely consider the transferability of trained policies to new tasks.

Graph Attention Multi-agent Reinforcement Learning +1

WATTNet: Learning to Trade FX via Hierarchical Spatio-Temporal Representation of Highly Multivariate Time Series

2 code implementations24 Sep 2019 Michael Poli, Jinkyoo Park, Ilija Ilievski

Finance is a particularly challenging application area for deep learning models due to low noise-to-signal ratio, non-stationarity, and partial observability.

Time Series Time Series Analysis

Port-Hamiltonian Approach to Neural Network Training

2 code implementations6 Sep 2019 Stefano Massaroli, Michael Poli, Federico Califano, Angela Faragasso, Jinkyoo Park, Atsushi Yamashita, Hajime Asama

Neural networks are discrete entities: subdivided into discrete layers and parametrized by weights which are iteratively optimized via difference equations.

Time Series Forecasting

Learning NP-Hard Multi-Agent Assignment Planning using GNN: Inference on a Random Graph and Provable Auction-Fitted Q-learning

no code implementations29 May 2019 Hyunwook Kang, Taehwan Kwon, Jinkyoo Park, James R. Morrison

In representing the MRRC problem as a sequential decision-making problem, we observe that each state can be represented as an extension of probabilistic graphical models (PGMs), which we refer to as random PGMs.

Combinatorial Optimization Decision Making +3

Multi-Agent Actor-Critic with Generative Cooperative Policy Network

no code implementations22 Oct 2018 Heechang Ryu, Hayong Shin, Jinkyoo Park

We propose an efficient multi-agent reinforcement learning approach to derive equilibrium strategies for multi-agents who are participating in a Markov game.

Multi-agent Reinforcement Learning Reinforcement Learning (RL)

Cannot find the paper you are looking for? You can Submit a new open access paper.