Search Results for author: Thomas Jiralerspong

Found 7 papers, 4 papers with code

Efficient Causal Graph Discovery Using Large Language Models

1 code implementation2 Feb 2024 Thomas Jiralerspong, Xiaoyin Chen, Yash More, Vedant Shah, Yoshua Bengio

We propose a novel framework that leverages LLMs for full causal graph discovery.

Network Analysis of the iNaturalist Citizen Science Community

no code implementations16 Oct 2023 Yu Lu Liu, Thomas Jiralerspong

We frame the data from iNaturalist as a bipartite network and use visualizations as well as established network science techniques to gain insights into the structure and interactions between users in citizen science projects.

Link Prediction

Forecaster: Towards Temporally Abstract Tree-Search Planning from Pixels

no code implementations16 Oct 2023 Thomas Jiralerspong, Flemming Kondrup, Doina Precup, Khimya Khetarpal

The ability to plan at many different levels of abstraction enables agents to envision the long-term repercussions of their decisions and thus enables sample-efficient learning.

Hierarchical Reinforcement Learning

Delta-AI: Local objectives for amortized inference in sparse graphical models

1 code implementation3 Oct 2023 Jean-Pierre Falet, Hae Beom Lee, Nikolay Malkin, Chen Sun, Dragos Secrieru, Thomas Jiralerspong, Dinghuai Zhang, Guillaume Lajoie, Yoshua Bengio

We present a new algorithm for amortized inference in sparse probabilistic graphical models (PGMs), which we call $\Delta$-amortized inference ($\Delta$-AI).

Contrastive Retrospection: honing in on critical steps for rapid learning and generalization in RL

1 code implementation NeurIPS 2023 Chen Sun, Wannan Yang, Thomas Jiralerspong, Dane Malenfant, Benjamin Alsbury-Nealy, Yoshua Bengio, Blake Richards

Distinct from other contemporary RL approaches to credit assignment, ConSpec takes advantage of the fact that it is easier to retrospectively identify the small set of steps that success is contingent upon (and ignoring other states) than it is to prospectively predict reward at every taken step.

Contrastive Learning Out-of-Distribution Generalization +1

Towards Safe Mechanical Ventilation Treatment Using Deep Offline Reinforcement Learning

1 code implementation5 Oct 2022 Flemming Kondrup, Thomas Jiralerspong, Elaine Lau, Nathan de Lara, Jacob Shkrob, My Duc Tran, Doina Precup, Sumana Basu

We design a clinically relevant intermediate reward that encourages continuous improvement of the patient vitals as well as addresses the challenge of sparse reward in RL.

Q-Learning reinforcement-learning +1

Cannot find the paper you are looking for? You can Submit a new open access paper.