Picture for Scott Niekum

Scott Niekum

Regularized Latent Dynamics Prediction is a Strong Baseline For Behavioral Foundation Models

Add code
Mar 16, 2026
Viaarxiv icon

Safe RLHF Beyond Expectation: Stochastic Dominance for Universal Spectral Risk Control

Add code
Mar 11, 2026
Viaarxiv icon

DeepFleet: Multi-Agent Foundation Models for Mobile Robots

Add code
Aug 12, 2025
Figure 1 for DeepFleet: Multi-Agent Foundation Models for Mobile Robots
Figure 2 for DeepFleet: Multi-Agent Foundation Models for Mobile Robots
Figure 3 for DeepFleet: Multi-Agent Foundation Models for Mobile Robots
Figure 4 for DeepFleet: Multi-Agent Foundation Models for Mobile Robots
Viaarxiv icon

Null Counterfactual Factor Interactions for Goal-Conditioned Reinforcement Learning

Add code
May 06, 2025
Figure 1 for Null Counterfactual Factor Interactions for Goal-Conditioned Reinforcement Learning
Figure 2 for Null Counterfactual Factor Interactions for Goal-Conditioned Reinforcement Learning
Figure 3 for Null Counterfactual Factor Interactions for Goal-Conditioned Reinforcement Learning
Figure 4 for Null Counterfactual Factor Interactions for Goal-Conditioned Reinforcement Learning
Viaarxiv icon

Pairwise or Pointwise? Evaluating Feedback Protocols for Bias in LLM-Based Evaluation

Add code
Apr 20, 2025
Figure 1 for Pairwise or Pointwise? Evaluating Feedback Protocols for Bias in LLM-Based Evaluation
Figure 2 for Pairwise or Pointwise? Evaluating Feedback Protocols for Bias in LLM-Based Evaluation
Figure 3 for Pairwise or Pointwise? Evaluating Feedback Protocols for Bias in LLM-Based Evaluation
Figure 4 for Pairwise or Pointwise? Evaluating Feedback Protocols for Bias in LLM-Based Evaluation
Viaarxiv icon

An Optimal Discriminator Weighted Imitation Perspective for Reinforcement Learning

Add code
Apr 17, 2025
Viaarxiv icon

Fast Adaptation with Behavioral Foundation Models

Add code
Apr 10, 2025
Viaarxiv icon

Supervised Reward Inference

Add code
Feb 25, 2025
Figure 1 for Supervised Reward Inference
Figure 2 for Supervised Reward Inference
Figure 3 for Supervised Reward Inference
Figure 4 for Supervised Reward Inference
Viaarxiv icon

Influencing Humans to Conform to Preference Models for RLHF

Add code
Jan 11, 2025
Figure 1 for Influencing Humans to Conform to Preference Models for RLHF
Figure 2 for Influencing Humans to Conform to Preference Models for RLHF
Figure 3 for Influencing Humans to Conform to Preference Models for RLHF
Figure 4 for Influencing Humans to Conform to Preference Models for RLHF
Viaarxiv icon

RL Zero: Zero-Shot Language to Behaviors without any Supervision

Add code
Dec 07, 2024
Viaarxiv icon