Picture for Alfredo Garcia

Alfredo Garcia

Joint Demonstration and Preference Learning Improves Policy Alignment with Human Feedback

Add code
Jun 11, 2024
Viaarxiv icon

Getting More Juice Out of the SFT Data: Reward Learning from Human Demonstration Improves SFT for LLM Alignment

Add code
May 29, 2024
Viaarxiv icon

Global Convergence of Decentralized Retraction-Free Optimization on the Stiefel Manifold

Add code
May 19, 2024
Viaarxiv icon

Regularized Q-Learning with Linear Function Approximation

Add code
Jan 26, 2024
Viaarxiv icon

Resolving uncertainty on the fly: Modeling adaptive driving behavior as active inference

Add code
Nov 10, 2023
Viaarxiv icon

A Unified View on Solving Objective Mismatch in Model-Based Reinforcement Learning

Add code
Oct 10, 2023
Figure 1 for A Unified View on Solving Objective Mismatch in Model-Based Reinforcement Learning
Viaarxiv icon

A Bayesian Approach to Robust Inverse Reinforcement Learning

Add code
Sep 15, 2023
Figure 1 for A Bayesian Approach to Robust Inverse Reinforcement Learning
Figure 2 for A Bayesian Approach to Robust Inverse Reinforcement Learning
Figure 3 for A Bayesian Approach to Robust Inverse Reinforcement Learning
Figure 4 for A Bayesian Approach to Robust Inverse Reinforcement Learning
Viaarxiv icon

An active inference model of car following: Advantages and applications

Add code
Mar 27, 2023
Figure 1 for An active inference model of car following: Advantages and applications
Figure 2 for An active inference model of car following: Advantages and applications
Figure 3 for An active inference model of car following: Advantages and applications
Figure 4 for An active inference model of car following: Advantages and applications
Viaarxiv icon

Understanding Expertise through Demonstrations: A Maximum Likelihood Framework for Offline Inverse Reinforcement Learning

Add code
Feb 15, 2023
Figure 1 for Understanding Expertise through Demonstrations: A Maximum Likelihood Framework for Offline Inverse Reinforcement Learning
Figure 2 for Understanding Expertise through Demonstrations: A Maximum Likelihood Framework for Offline Inverse Reinforcement Learning
Figure 3 for Understanding Expertise through Demonstrations: A Maximum Likelihood Framework for Offline Inverse Reinforcement Learning
Figure 4 for Understanding Expertise through Demonstrations: A Maximum Likelihood Framework for Offline Inverse Reinforcement Learning
Viaarxiv icon

Maximum-Likelihood Inverse Reinforcement Learning with Finite-Time Guarantees

Add code
Oct 04, 2022
Figure 1 for Maximum-Likelihood Inverse Reinforcement Learning with Finite-Time Guarantees
Figure 2 for Maximum-Likelihood Inverse Reinforcement Learning with Finite-Time Guarantees
Figure 3 for Maximum-Likelihood Inverse Reinforcement Learning with Finite-Time Guarantees
Viaarxiv icon