Alert button
Picture for Noam Brown

Noam Brown

Alert button

The Update Equivalence Framework for Decision-Time Planning

Apr 25, 2023
Samuel Sokota, Gabriele Farina, David J. Wu, Hengyuan Hu, Kevin A. Wang, J. Zico Kolter, Noam Brown

Figure 1 for The Update Equivalence Framework for Decision-Time Planning
Figure 2 for The Update Equivalence Framework for Decision-Time Planning
Figure 3 for The Update Equivalence Framework for Decision-Time Planning
Figure 4 for The Update Equivalence Framework for Decision-Time Planning
Viaarxiv icon

Abstracting Imperfect Information Away from Two-Player Zero-Sum Games

Jan 22, 2023
Samuel Sokota, Ryan D'Orazio, Chun Kai Ling, David J. Wu, J. Zico Kolter, Noam Brown

Figure 1 for Abstracting Imperfect Information Away from Two-Player Zero-Sum Games
Figure 2 for Abstracting Imperfect Information Away from Two-Player Zero-Sum Games
Figure 3 for Abstracting Imperfect Information Away from Two-Player Zero-Sum Games
Figure 4 for Abstracting Imperfect Information Away from Two-Player Zero-Sum Games
Viaarxiv icon

Mastering the Game of No-Press Diplomacy via Human-Regularized Reinforcement Learning and Planning

Oct 11, 2022
Anton Bakhtin, David J Wu, Adam Lerer, Jonathan Gray, Athul Paul Jacob, Gabriele Farina, Alexander H Miller, Noam Brown

Figure 1 for Mastering the Game of No-Press Diplomacy via Human-Regularized Reinforcement Learning and Planning
Figure 2 for Mastering the Game of No-Press Diplomacy via Human-Regularized Reinforcement Learning and Planning
Figure 3 for Mastering the Game of No-Press Diplomacy via Human-Regularized Reinforcement Learning and Planning
Figure 4 for Mastering the Game of No-Press Diplomacy via Human-Regularized Reinforcement Learning and Planning
Viaarxiv icon

Human-AI Coordination via Human-Regularized Search and Learning

Oct 11, 2022
Hengyuan Hu, David J Wu, Adam Lerer, Jakob Foerster, Noam Brown

Figure 1 for Human-AI Coordination via Human-Regularized Search and Learning
Figure 2 for Human-AI Coordination via Human-Regularized Search and Learning
Figure 3 for Human-AI Coordination via Human-Regularized Search and Learning
Viaarxiv icon

A Unified Approach to Reinforcement Learning, Quantal Response Equilibria, and Two-Player Zero-Sum Games

Jun 12, 2022
Samuel Sokota, Ryan D'Orazio, J. Zico Kolter, Nicolas Loizou, Marc Lanctot, Ioannis Mitliagkas, Noam Brown, Christian Kroer

Figure 1 for A Unified Approach to Reinforcement Learning, Quantal Response Equilibria, and Two-Player Zero-Sum Games
Figure 2 for A Unified Approach to Reinforcement Learning, Quantal Response Equilibria, and Two-Player Zero-Sum Games
Figure 3 for A Unified Approach to Reinforcement Learning, Quantal Response Equilibria, and Two-Player Zero-Sum Games
Figure 4 for A Unified Approach to Reinforcement Learning, Quantal Response Equilibria, and Two-Player Zero-Sum Games
Viaarxiv icon

Modeling Strong and Human-Like Gameplay with KL-Regularized Search

Dec 14, 2021
Athul Paul Jacob, David J. Wu, Gabriele Farina, Adam Lerer, Anton Bakhtin, Jacob Andreas, Noam Brown

Figure 1 for Modeling Strong and Human-Like Gameplay with KL-Regularized Search
Figure 2 for Modeling Strong and Human-Like Gameplay with KL-Regularized Search
Figure 3 for Modeling Strong and Human-Like Gameplay with KL-Regularized Search
Figure 4 for Modeling Strong and Human-Like Gameplay with KL-Regularized Search
Viaarxiv icon

No-Press Diplomacy from Scratch

Oct 06, 2021
Anton Bakhtin, David Wu, Adam Lerer, Noam Brown

Figure 1 for No-Press Diplomacy from Scratch
Figure 2 for No-Press Diplomacy from Scratch
Figure 3 for No-Press Diplomacy from Scratch
Figure 4 for No-Press Diplomacy from Scratch
Viaarxiv icon

Scalable Online Planning via Reinforcement Learning Fine-Tuning

Sep 30, 2021
Arnaud Fickinger, Hengyuan Hu, Brandon Amos, Stuart Russell, Noam Brown

Figure 1 for Scalable Online Planning via Reinforcement Learning Fine-Tuning
Figure 2 for Scalable Online Planning via Reinforcement Learning Fine-Tuning
Figure 3 for Scalable Online Planning via Reinforcement Learning Fine-Tuning
Figure 4 for Scalable Online Planning via Reinforcement Learning Fine-Tuning
Viaarxiv icon

Learned Belief Search: Efficiently Improving Policies in Partially Observable Settings

Jun 16, 2021
Hengyuan Hu, Adam Lerer, Noam Brown, Jakob Foerster

Figure 1 for Learned Belief Search: Efficiently Improving Policies in Partially Observable Settings
Figure 2 for Learned Belief Search: Efficiently Improving Policies in Partially Observable Settings
Figure 3 for Learned Belief Search: Efficiently Improving Policies in Partially Observable Settings
Figure 4 for Learned Belief Search: Efficiently Improving Policies in Partially Observable Settings
Viaarxiv icon

Off-Belief Learning

Mar 06, 2021
Hengyuan Hu, Adam Lerer, Brandon Cui, Luis Pineda, David Wu, Noam Brown, Jakob Foerster

Figure 1 for Off-Belief Learning
Figure 2 for Off-Belief Learning
Figure 3 for Off-Belief Learning
Figure 4 for Off-Belief Learning
Viaarxiv icon