Picture for Sarath Chandar

Sarath Chandar

Language Model-In-The-Loop: Data Optimal Approach to Learn-To-Recommend Actions in Text Games

Add code
Nov 13, 2023
Figure 1 for Language Model-In-The-Loop: Data Optimal Approach to Learn-To-Recommend Actions in Text Games
Figure 2 for Language Model-In-The-Loop: Data Optimal Approach to Learn-To-Recommend Actions in Text Games
Figure 3 for Language Model-In-The-Loop: Data Optimal Approach to Learn-To-Recommend Actions in Text Games
Figure 4 for Language Model-In-The-Loop: Data Optimal Approach to Learn-To-Recommend Actions in Text Games
Viaarxiv icon

Self-Influence Guided Data Reweighting for Language Model Pre-training

Add code
Nov 02, 2023
Figure 1 for Self-Influence Guided Data Reweighting for Language Model Pre-training
Figure 2 for Self-Influence Guided Data Reweighting for Language Model Pre-training
Figure 3 for Self-Influence Guided Data Reweighting for Language Model Pre-training
Figure 4 for Self-Influence Guided Data Reweighting for Language Model Pre-training
Viaarxiv icon

EpiK-Eval: Evaluation for Language Models as Epistemic Models

Add code
Oct 23, 2023
Viaarxiv icon

Faithfulness Measurable Masked Language Models

Add code
Oct 11, 2023
Viaarxiv icon

Towards Few-shot Coordination: Revisiting Ad-hoc Teamplay Challenge In the Game of Hanabi

Add code
Aug 20, 2023
Figure 1 for Towards Few-shot Coordination: Revisiting Ad-hoc Teamplay Challenge In the Game of Hanabi
Figure 2 for Towards Few-shot Coordination: Revisiting Ad-hoc Teamplay Challenge In the Game of Hanabi
Figure 3 for Towards Few-shot Coordination: Revisiting Ad-hoc Teamplay Challenge In the Game of Hanabi
Figure 4 for Towards Few-shot Coordination: Revisiting Ad-hoc Teamplay Challenge In the Game of Hanabi
Viaarxiv icon

Lookbehind Optimizer: k steps back, 1 step forward

Add code
Jul 31, 2023
Viaarxiv icon

Promoting Exploration in Memory-Augmented Adam using Critical Momenta

Add code
Jul 18, 2023
Viaarxiv icon

Thompson sampling for improved exploration in GFlowNets

Add code
Jun 30, 2023
Viaarxiv icon

Measuring the Knowledge Acquisition-Utilization Gap in Pretrained Language Models

Add code
May 24, 2023
Viaarxiv icon

Should We Attend More or Less? Modulating Attention for Fairness

Add code
May 22, 2023
Figure 1 for Should We Attend More or Less? Modulating Attention for Fairness
Figure 2 for Should We Attend More or Less? Modulating Attention for Fairness
Figure 3 for Should We Attend More or Less? Modulating Attention for Fairness
Figure 4 for Should We Attend More or Less? Modulating Attention for Fairness
Viaarxiv icon