no code implementations • 20 Jan 2023 • Maxime Bouton, Jaeseong Jeong, Jose Outes, Adriano Mendo, Alexandros Nikou
Future generations of mobile networks are expected to contain more and more antennas with growing complexity and more parameters.
Multi-agent Reinforcement Learning reinforcement-learning +1
no code implementations • 6 Jan 2022 • Filippo Vannella, Alexandre Proutiere, Yassir Jedra, Jaeseong Jeong
In this paper, we devise algorithms learning optimal tilt control policies from existing data (in the so-called passive learning setting) or from data actively generated by the algorithms (the active learning setting).
no code implementations • 27 Dec 2021 • Yifei Jin, Filippo Vannella, Maxime Bouton, Jaeseong Jeong, Ezeddin Al Hakim
GAQ relies on a graph attention mechanism to select relevant neighbors information, improve the agent state representation, and update the tilt control policy based on a history of observations using a Deep Q-Network (DQN).
no code implementations • 10 Sep 2021 • Heunchul Lee, Jaeseong Jeong
A multi-agent deep reinforcement learning (MADRL) is a promising approach to challenging problems in wireless environments involving multiple decision-makers (or actors) with high-dimensional continuous action space.
no code implementations • 30 Jun 2020 • Heunchul Lee, Maksym Girnyk, Jaeseong Jeong
To demonstrate the optimality of the proposed DRL-based precoding framework, we explicitly consider a simple MIMO environment for which the optimal solution can be obtained analytically and show that DQN- and DDPG-based agents can learn the near-optimal policy to map the environment state of MIMO system to a precoder that maximizes the reward function, respectively, in the codebook-based and non-codebook based MIMO precoding systems.
no code implementations • 21 May 2020 • Filippo Vannella, Jaeseong Jeong, Alexandre Proutiere
In this paper, we circumvent these issues by learning an improved policy in an offline manner using existing data collected on real networks.
no code implementations • 12 Jul 2015 • Jaeseong Jeong, Mathieu Leconte, Alexandre Proutiere
In this paper, we develop cluster-aided predictors that exploit past trajectories collected from all users to predict the next location of a given user.