1 code implementation • EMNLP 2021 • Subhajit Chaudhury, Prithviraj Sen, Masaki Ono, Daiki Kimura, Michiaki Tatsubori, Asim Munawar
We outline a method for end-to-end differentiable symbolic rule learning and show that such symbolic policies outperform previous state-of-the-art methods in text-based RL for the coin collector environment from 5-10x fewer training games.
no code implementations • 31 May 2023 • Maxwell Crouse, Ramon Astudillo, Tahira Naseem, Subhajit Chaudhury, Pavan Kapanipathi, Salim Roukos, Alexander Gray
We introduce Logical Offline Cycle Consistency Optimization (LOCCO), a scalable, semi-supervised method for training a neural semantic parser.
no code implementations • 7 May 2023 • Maxwell Crouse, Pavan Kapanipathi, Subhajit Chaudhury, Tahira Naseem, Ramon Astudillo, Achille Fokoue, Tim Klinger
Nearly all general-purpose neural semantic parsers generate logical forms in a strictly top-down autoregressive fashion.
no code implementations • 23 Oct 2022 • Heshan Fernando, Han Shen, Miao Liu, Subhajit Chaudhury, Keerthiram Murugesan, Tianyi Chen
Machine learning problems with multiple objective functions appear either in learning with multiple criteria where learning has to make a trade-off between multiple performance metrics such as fairness, safety and accuracy; or, in multi-task learning where multiple tasks are optimized jointly, sharing inductive bias between them.
no code implementations • EMNLP 2021 • Daiki Kimura, Masaki Ono, Subhajit Chaudhury, Ryosuke Kohita, Akifumi Wachi, Don Joven Agravante, Michiaki Tatsubori, Asim Munawar, Alexander Gray
Deep reinforcement learning (RL) methods often require many trials before convergence, and no direct interpretability of trained policies is provided.
1 code implementation • ACL 2021 • Daiki Kimura, Subhajit Chaudhury, Masaki Ono, Michiaki Tatsubori, Don Joven Agravante, Asim Munawar, Akifumi Wachi, Ryosuke Kohita, Alexander Gray
We present Logical Optimal Actions (LOA), an action decision architecture of reinforcement learning applications with a neuro-symbolic framework which is a combination of neural network and symbolic knowledge acquisition approach for natural language interaction games.
no code implementations • 9 Jun 2021 • Keerthiram Murugesan, Subhajit Chaudhury, Kartik Talamadupula
This improves the agent's overall understanding of the game 'scene' and objects' relationships to the world around them, and the variety of visual representations on offer allow the agent to generate a better generalization of a relationship.
no code implementations • 3 Mar 2021 • Daiki Kimura, Subhajit Chaudhury, Akifumi Wachi, Ryosuke Kohita, Asim Munawar, Michiaki Tatsubori, Alexander Gray
Specifically, we propose an integrated method that enables model-free reinforcement learning from external knowledge sources in an LNNs-based logical constrained framework such as action shielding and guide.
1 code implementation • 3 Dec 2020 • Hiya Roy, Subhajit Chaudhury, Toshihiko Yamasaki, Tatsuaki Hashimoto
To alleviate these problems, we investigate if it is possible to obtain better performance by training the networks using frequency domain information (Discrete Fourier Transform) along with the spatial domain information.
no code implementations • 26 Oct 2020 • Thomas Carta, Subhajit Chaudhury, Kartik Talamadupula, Michiaki Tatsubori
The goal is to force an RL agent to use both text and visual features to predict natural language action commands for solving the final task of cooking a meal.
1 code implementation • EMNLP 2020 • Subhajit Chaudhury, Daiki Kimura, Kartik Talamadupula, Michiaki Tatsubori, Asim Munawar, Ryuki Tachibana
Our bootstrapped agent shows improved generalization in solving unseen TextWorld games, using 10x-20x fewer training games compared to previous state-of-the-art methods despite requiring less number of training episodes.
1 code implementation • 14 Mar 2020 • Subhajit Chaudhury, Toshihiko Yamasaki
In this paper, we study the generalization properties of neural networks under input perturbations and show that minimal training data corruption by a few pixel modifications can cause drastic overfitting.
no code implementations • 19 Feb 2020 • Subhajit Chaudhury, Daiki Kimura, Phongtharin Vinayavekhin, Asim Munawar, Ryuki Tachibana, Koji Ito, Yuki Inaba, Minoru Matsumoto, Shuji Kidokoro, Hiroki Ozaki
In this paper, we study the case of event detection in sports videos for unstructured environments with arbitrary camera angles.
no code implementations • 15 Jan 2020 • Sourav Mishra, Subhajit Chaudhury, Hideaki Imaizumi, Toshihiko Yamasaki
This paper aims to evaluate the suitability of current deep learning methods for clinical workflow especially by focusing on dermatology.
no code implementations • 14 Apr 2019 • Hiya Roy, Subhajit Chaudhury, Toshihiko Yamasaki, Danielle DeLatte, Makiko Ohtake, Tatsuaki Hashimoto
Image restoration is a technique that reconstructs a feasible estimate of the original image from the noisy observation.
no code implementations • 2 Oct 2018 • Subhajit Chaudhury, Daiki Kimura, Asim Munawar, Ryuki Tachibana
Experimental results show that the proposed adversarial learning method from raw videos produces a similar performance to state-of-the-art imitation learning techniques while frequently outperforming existing hand-crafted video imitation methods.
1 code implementation • 21 Sep 2018 • Tu-Hoa Pham, Giovanni De Magistris, Don Joven Agravante, Subhajit Chaudhury, Asim Munawar, Ryuki Tachibana
We consider the problem of reinforcement learning under safety requirements, in which an agent is trained to complete a given task, typically formalized as the maximization of a reward signal over time, while concurrently avoiding undesirable actions or states, associated to lower rewards, or penalties.
no code implementations • 4 Jul 2018 • Tadanobu Inoue, Subhajit Chaudhury, Giovanni De Magistris, Sakyasingha Dasgupta
Capturing and labeling camera images in the real world is an expensive task, whereas synthesizing labeled images in a simulation environment is easy for collecting large-scale image data.
no code implementations • 22 Jun 2018 • Phongtharin Vinayavekhin, Subhajit Chaudhury, Asim Munawar, Don Joven Agravante, Giovanni De Magistris, Daiki Kimura, Ryuki Tachibana
This paper is a contribution towards interpretability of the deep learning models in different applications of time-series.
no code implementations • 2 Jun 2018 • Daiki Kimura, Subhajit Chaudhury, Ryuki Tachibana, Sakyasingha Dasgupta
During reinforcement learning, the agent predicts the reward as a function of the difference between the actual state and the state predicted by the internal model.
no code implementations • ICLR 2018 • Daiki Kimura, Subhajit Chaudhury, Ryuki Tachibana, Sakyasingha Dasgupta
We present a novel reward estimation method that is based on a finite sample of optimal state trajectories from expert demon- strations and can be used for guiding an agent to mimic the expert behavior.
no code implementations • ICLR 2018 • Subhajit Chaudhury, Daiki Kimura, Tadanobu Inoue, Ryuki Tachibana
We present a model-based imitation learning method that can learn environment-specific optimal actions only from expert state trajectories.
no code implementations • 20 Sep 2017 • Tadanobu Inoue, Subhajit Chaudhury, Giovanni De Magistris, Sakyasingha Dasgupta
It detects object positions 6 to 7 times more precisely than the baseline of directly learning from the dataset of the real images.
no code implementations • 4 Jul 2017 • Subhajit Chaudhury, Sakyasingha Dasgupta, Asim Munawar, Md. A. Salam Khan, Ryuki Tachibana
We present a conditional generative model that maps low-dimensional embeddings of multiple modalities of data to a common latent space hence extracting semantic relationships between them.
no code implementations • 14 Nov 2016 • Subhajit Chaudhury, Hiya Roy
We propose a fully convolutional model, that learns a direct end-to-end mapping between the corrupted images as input and the desired clean images as output.