Picture for Shlomo Zilberstein

Shlomo Zilberstein

University of Massachuetts Amherst

Terrarium: Revisiting the Blackboard for Multi-Agent Safety, Privacy, and Security Studies

Add code
Oct 16, 2025
Viaarxiv icon

Aligning LLMs on a Budget: Inference-Time Alignment with Heuristic Reward Models

Add code
Aug 07, 2025
Figure 1 for Aligning LLMs on a Budget: Inference-Time Alignment with Heuristic Reward Models
Figure 2 for Aligning LLMs on a Budget: Inference-Time Alignment with Heuristic Reward Models
Figure 3 for Aligning LLMs on a Budget: Inference-Time Alignment with Heuristic Reward Models
Figure 4 for Aligning LLMs on a Budget: Inference-Time Alignment with Heuristic Reward Models
Viaarxiv icon

Proceedings of 1st Workshop on Advancing Artificial Intelligence through Theory of Mind

Add code
Apr 28, 2025
Figure 1 for Proceedings of 1st Workshop on Advancing Artificial Intelligence through Theory of Mind
Figure 2 for Proceedings of 1st Workshop on Advancing Artificial Intelligence through Theory of Mind
Figure 3 for Proceedings of 1st Workshop on Advancing Artificial Intelligence through Theory of Mind
Figure 4 for Proceedings of 1st Workshop on Advancing Artificial Intelligence through Theory of Mind
Viaarxiv icon

Distributed Multi-Agent Coordination Using Multi-Modal Foundation Models

Add code
Jan 24, 2025
Viaarxiv icon

MAPLE: A Framework for Active Preference Learning Guided by Large Language Models

Add code
Dec 10, 2024
Figure 1 for MAPLE: A Framework for Active Preference Learning Guided by Large Language Models
Figure 2 for MAPLE: A Framework for Active Preference Learning Guided by Large Language Models
Figure 3 for MAPLE: A Framework for Active Preference Learning Guided by Large Language Models
Figure 4 for MAPLE: A Framework for Active Preference Learning Guided by Large Language Models
Viaarxiv icon

RL$^3$: Boosting Meta Reinforcement Learning via RL inside RL$^2$

Add code
Jun 28, 2023
Figure 1 for RL$^3$: Boosting Meta Reinforcement Learning via RL inside RL$^2$
Figure 2 for RL$^3$: Boosting Meta Reinforcement Learning via RL inside RL$^2$
Figure 3 for RL$^3$: Boosting Meta Reinforcement Learning via RL inside RL$^2$
Figure 4 for RL$^3$: Boosting Meta Reinforcement Learning via RL inside RL$^2$
Viaarxiv icon

Adaptive Rollout Length for Model-Based RL Using Model-Free Deep RL

Add code
Jun 07, 2022
Figure 1 for Adaptive Rollout Length for Model-Based RL Using Model-Free Deep RL
Figure 2 for Adaptive Rollout Length for Model-Based RL Using Model-Free Deep RL
Viaarxiv icon

Dense Crowd Flow-Informed Path Planning

Add code
Jun 01, 2022
Figure 1 for Dense Crowd Flow-Informed Path Planning
Figure 2 for Dense Crowd Flow-Informed Path Planning
Figure 3 for Dense Crowd Flow-Informed Path Planning
Figure 4 for Dense Crowd Flow-Informed Path Planning
Viaarxiv icon

A Unifying Framework for Causal Explanation of Sequential Decision Making

Add code
May 30, 2022
Figure 1 for A Unifying Framework for Causal Explanation of Sequential Decision Making
Figure 2 for A Unifying Framework for Causal Explanation of Sequential Decision Making
Figure 3 for A Unifying Framework for Causal Explanation of Sequential Decision Making
Figure 4 for A Unifying Framework for Causal Explanation of Sequential Decision Making
Viaarxiv icon

Competence-Aware Path Planning via Introspective Perception

Add code
Sep 28, 2021
Figure 1 for Competence-Aware Path Planning via Introspective Perception
Figure 2 for Competence-Aware Path Planning via Introspective Perception
Figure 3 for Competence-Aware Path Planning via Introspective Perception
Figure 4 for Competence-Aware Path Planning via Introspective Perception
Viaarxiv icon