Search Results for author: Yash Satsangi

Found 9 papers, 1 papers with code

Estimating Class Separability of Datasets Using Persistent Homology with Application to LLM Fine-Tuning

no code implementations24 May 2023 Najah Ghalyan, Kostis Gourgoulias, Yash Satsangi, Sean Moran, Maxime Labonne, Joseph Sabelja

This paper proposes a method to estimate the class separability of an unlabeled text dataset by inspecting the topological characteristics of sentence-transformer embeddings of the text.

Language Modelling Sentence +1

Learning to Be Cautious

no code implementations29 Oct 2021 Montaser Mohammedalamen, Dustin Morrill, Alexander Sieusahai, Yash Satsangi, Michael Bowling

An agent that could learn to be cautious would overcome this challenge by discovering for itself when and how to behave cautiously.

counterfactual

Useful Policy Invariant Shaping from Arbitrary Advice

no code implementations2 Nov 2020 Paniz Behboudian, Yash Satsangi, Matthew E. Taylor, Anna Harutyunyan, Michael Bowling

Furthermore, if the reward is constructed from a potential function, the optimal policy is guaranteed to be unaltered.

Exploiting Submodular Value Functions For Scaling Up Active Perception

no code implementations21 Sep 2020 Yash Satsangi, Shimon Whiteson, Frans A. Oliehoek, Matthijs T. J. Spaan

Furthermore, we show that, under certain conditions, including submodularity, the value function computed using greedy PBVI is guaranteed to have bounded error with respect to the optimal value function.

Maximizing Information Gain in Partially Observable Environments via Prediction Reward

no code implementations11 May 2020 Yash Satsangi, Sungsu Lim, Shimon Whiteson, Frans Oliehoek, Martha White

Information gathering in a partially observable environment can be formulated as a reinforcement learning (RL), problem where the reward depends on the agent's uncertainty.

Question Answering Reinforcement Learning (RL)

Cannot find the paper you are looking for? You can Submit a new open access paper.