Alert button
Picture for Yash Satsangi

Yash Satsangi

Alert button

An Unsupervised Method for Estimating Class Separability of Datasets with Application to LLMs Fine-Tuning

Add code
Bookmark button
Alert button
May 24, 2023
Najah Ghalyan, Kostis Gourgoulias, Yash Satsangi, Sean Moran, Maxime Labonne, Joseph Sabelja

Figure 1 for An Unsupervised Method for Estimating Class Separability of Datasets with Application to LLMs Fine-Tuning
Figure 2 for An Unsupervised Method for Estimating Class Separability of Datasets with Application to LLMs Fine-Tuning
Figure 3 for An Unsupervised Method for Estimating Class Separability of Datasets with Application to LLMs Fine-Tuning
Figure 4 for An Unsupervised Method for Estimating Class Separability of Datasets with Application to LLMs Fine-Tuning
Viaarxiv icon

Bandit-Based Policy Invariant Explicit Shaping for Incorporating External Advice in Reinforcement Learning

Add code
Bookmark button
Alert button
Apr 14, 2023
Yash Satsangi, Paniz Behboudian

Figure 1 for Bandit-Based Policy Invariant Explicit Shaping for Incorporating External Advice in Reinforcement Learning
Figure 2 for Bandit-Based Policy Invariant Explicit Shaping for Incorporating External Advice in Reinforcement Learning
Figure 3 for Bandit-Based Policy Invariant Explicit Shaping for Incorporating External Advice in Reinforcement Learning
Viaarxiv icon

Topical: Learning Repository Embeddings from Source Code using Attention

Add code
Bookmark button
Alert button
Aug 19, 2022
Agathe Lherondelle, Yash Satsangi, Fran Silavong, Shaltiel Eloul, Sean Moran

Figure 1 for Topical: Learning Repository Embeddings from Source Code using Attention
Figure 2 for Topical: Learning Repository Embeddings from Source Code using Attention
Figure 3 for Topical: Learning Repository Embeddings from Source Code using Attention
Figure 4 for Topical: Learning Repository Embeddings from Source Code using Attention
Viaarxiv icon

Learning to Be Cautious

Add code
Bookmark button
Alert button
Oct 29, 2021
Montaser Mohammedalamen, Dustin Morrill, Alexander Sieusahai, Yash Satsangi, Michael Bowling

Figure 1 for Learning to Be Cautious
Figure 2 for Learning to Be Cautious
Figure 3 for Learning to Be Cautious
Figure 4 for Learning to Be Cautious
Viaarxiv icon

Useful Policy Invariant Shaping from Arbitrary Advice

Add code
Bookmark button
Alert button
Nov 02, 2020
Paniz Behboudian, Yash Satsangi, Matthew E. Taylor, Anna Harutyunyan, Michael Bowling

Figure 1 for Useful Policy Invariant Shaping from Arbitrary Advice
Figure 2 for Useful Policy Invariant Shaping from Arbitrary Advice
Figure 3 for Useful Policy Invariant Shaping from Arbitrary Advice
Figure 4 for Useful Policy Invariant Shaping from Arbitrary Advice
Viaarxiv icon

Exploiting Submodular Value Functions For Scaling Up Active Perception

Add code
Bookmark button
Alert button
Sep 21, 2020
Yash Satsangi, Shimon Whiteson, Frans A. Oliehoek, Matthijs T. J. Spaan

Figure 1 for Exploiting Submodular Value Functions For Scaling Up Active Perception
Figure 2 for Exploiting Submodular Value Functions For Scaling Up Active Perception
Figure 3 for Exploiting Submodular Value Functions For Scaling Up Active Perception
Figure 4 for Exploiting Submodular Value Functions For Scaling Up Active Perception
Viaarxiv icon

Real-Time Resource Allocation for Tracking Systems

Add code
Bookmark button
Alert button
Sep 21, 2020
Yash Satsangi, Shimon Whiteson, Frans A. Oliehoek, Henri Bouma

Figure 1 for Real-Time Resource Allocation for Tracking Systems
Figure 2 for Real-Time Resource Allocation for Tracking Systems
Figure 3 for Real-Time Resource Allocation for Tracking Systems
Figure 4 for Real-Time Resource Allocation for Tracking Systems
Viaarxiv icon

Maximizing Information Gain in Partially Observable Environments via Prediction Reward

Add code
Bookmark button
Alert button
May 11, 2020
Yash Satsangi, Sungsu Lim, Shimon Whiteson, Frans Oliehoek, Martha White

Figure 1 for Maximizing Information Gain in Partially Observable Environments via Prediction Reward
Figure 2 for Maximizing Information Gain in Partially Observable Environments via Prediction Reward
Figure 3 for Maximizing Information Gain in Partially Observable Environments via Prediction Reward
Figure 4 for Maximizing Information Gain in Partially Observable Environments via Prediction Reward
Viaarxiv icon

Probably Approximately Correct Greedy Maximization

Add code
Bookmark button
Alert button
Feb 25, 2016
Yash Satsangi, Shimon Whiteson, Frans A. Oliehoek

Figure 1 for Probably Approximately Correct Greedy Maximization
Viaarxiv icon