Alert button
Picture for Nicholay Topin

Nicholay Topin

Alert button

Use-Case-Grounded Simulations for Explanation Evaluation

Jun 05, 2022
Valerie Chen, Nari Johnson, Nicholay Topin, Gregory Plumb, Ameet Talwalkar

Figure 1 for Use-Case-Grounded Simulations for Explanation Evaluation
Figure 2 for Use-Case-Grounded Simulations for Explanation Evaluation
Figure 3 for Use-Case-Grounded Simulations for Explanation Evaluation
Figure 4 for Use-Case-Grounded Simulations for Explanation Evaluation
Viaarxiv icon

MAVIPER: Learning Decision Tree Policies for Interpretable Multi-Agent Reinforcement Learning

May 25, 2022
Stephanie Milani, Zhicheng Zhang, Nicholay Topin, Zheyuan Ryan Shi, Charles Kamhoua, Evangelos E. Papalexakis, Fei Fang

Figure 1 for MAVIPER: Learning Decision Tree Policies for Interpretable Multi-Agent Reinforcement Learning
Figure 2 for MAVIPER: Learning Decision Tree Policies for Interpretable Multi-Agent Reinforcement Learning
Figure 3 for MAVIPER: Learning Decision Tree Policies for Interpretable Multi-Agent Reinforcement Learning
Figure 4 for MAVIPER: Learning Decision Tree Policies for Interpretable Multi-Agent Reinforcement Learning
Viaarxiv icon

MineRL Diamond 2021 Competition: Overview, Results, and Lessons Learned

Feb 17, 2022
Anssi Kanervisto, Stephanie Milani, Karolis Ramanauskas, Nicholay Topin, Zichuan Lin, Junyou Li, Jianing Shi, Deheng Ye, Qiang Fu, Wei Yang, Weijun Hong, Zhongyue Huang, Haicheng Chen, Guangjun Zeng, Yue Lin, Vincent Micheli, Eloi Alonso, François Fleuret, Alexander Nikulin, Yury Belousov, Oleg Svidchenko, Aleksei Shpilman

Figure 1 for MineRL Diamond 2021 Competition: Overview, Results, and Lessons Learned
Figure 2 for MineRL Diamond 2021 Competition: Overview, Results, and Lessons Learned
Figure 3 for MineRL Diamond 2021 Competition: Overview, Results, and Lessons Learned
Figure 4 for MineRL Diamond 2021 Competition: Overview, Results, and Lessons Learned
Viaarxiv icon

A Survey of Explainable Reinforcement Learning

Feb 17, 2022
Stephanie Milani, Nicholay Topin, Manuela Veloso, Fei Fang

Figure 1 for A Survey of Explainable Reinforcement Learning
Figure 2 for A Survey of Explainable Reinforcement Learning
Figure 3 for A Survey of Explainable Reinforcement Learning
Figure 4 for A Survey of Explainable Reinforcement Learning
Viaarxiv icon

The MineRL BASALT Competition on Learning from Human Feedback

Jul 05, 2021
Rohin Shah, Cody Wild, Steven H. Wang, Neel Alex, Brandon Houghton, William Guss, Sharada Mohanty, Anssi Kanervisto, Stephanie Milani, Nicholay Topin, Pieter Abbeel, Stuart Russell, Anca Dragan

Figure 1 for The MineRL BASALT Competition on Learning from Human Feedback
Figure 2 for The MineRL BASALT Competition on Learning from Human Feedback
Viaarxiv icon

Towards robust and domain agnostic reinforcement learning competitions

Jun 07, 2021
William Hebgen Guss, Stephanie Milani, Nicholay Topin, Brandon Houghton, Sharada Mohanty, Andrew Melnik, Augustin Harter, Benoit Buschmaas, Bjarne Jaster, Christoph Berganski, Dennis Heitkamp, Marko Henning, Helge Ritter, Chengjie Wu, Xiaotian Hao, Yiming Lu, Hangyu Mao, Yihuan Mao, Chao Wang, Michal Opanowicz, Anssi Kanervisto, Yanick Schraner, Christian Scheller, Xiren Zhou, Lu Liu, Daichi Nishio, Toi Tsuneda, Karolis Ramanauskas, Gabija Juceviciute

Figure 1 for Towards robust and domain agnostic reinforcement learning competitions
Figure 2 for Towards robust and domain agnostic reinforcement learning competitions
Figure 3 for Towards robust and domain agnostic reinforcement learning competitions
Figure 4 for Towards robust and domain agnostic reinforcement learning competitions
Viaarxiv icon

Iterative Bounding MDPs: Learning Interpretable Policies via Non-Interpretable Methods

Feb 25, 2021
Nicholay Topin, Stephanie Milani, Fei Fang, Manuela Veloso

Figure 1 for Iterative Bounding MDPs: Learning Interpretable Policies via Non-Interpretable Methods
Figure 2 for Iterative Bounding MDPs: Learning Interpretable Policies via Non-Interpretable Methods
Figure 3 for Iterative Bounding MDPs: Learning Interpretable Policies via Non-Interpretable Methods
Figure 4 for Iterative Bounding MDPs: Learning Interpretable Policies via Non-Interpretable Methods
Viaarxiv icon

The MineRL 2020 Competition on Sample Efficient Reinforcement Learning using Human Priors

Jan 26, 2021
William H. Guss, Mario Ynocente Castro, Sam Devlin, Brandon Houghton, Noboru Sean Kuno, Crissman Loomis, Stephanie Milani, Sharada Mohanty, Keisuke Nakata, Ruslan Salakhutdinov, John Schulman, Shinya Shiroshita, Nicholay Topin, Avinash Ummadisingu, Oriol Vinyals

Figure 1 for The MineRL 2020 Competition on Sample Efficient Reinforcement Learning using Human Priors
Figure 2 for The MineRL 2020 Competition on Sample Efficient Reinforcement Learning using Human Priors
Figure 3 for The MineRL 2020 Competition on Sample Efficient Reinforcement Learning using Human Priors
Figure 4 for The MineRL 2020 Competition on Sample Efficient Reinforcement Learning using Human Priors
Viaarxiv icon

Guaranteeing Reproducibility in Deep Learning Competitions

May 12, 2020
Brandon Houghton, Stephanie Milani, Nicholay Topin, William Guss, Katja Hofmann, Diego Perez-Liebana, Manuela Veloso, Ruslan Salakhutdinov

Viaarxiv icon

Retrospective Analysis of the 2019 MineRL Competition on Sample Efficient Reinforcement Learning

Mar 27, 2020
Stephanie Milani, Nicholay Topin, Brandon Houghton, William H. Guss, Sharada P. Mohanty, Keisuke Nakata, Oriol Vinyals, Noboru Sean Kuno

Figure 1 for Retrospective Analysis of the 2019 MineRL Competition on Sample Efficient Reinforcement Learning
Figure 2 for Retrospective Analysis of the 2019 MineRL Competition on Sample Efficient Reinforcement Learning
Figure 3 for Retrospective Analysis of the 2019 MineRL Competition on Sample Efficient Reinforcement Learning
Figure 4 for Retrospective Analysis of the 2019 MineRL Competition on Sample Efficient Reinforcement Learning
Viaarxiv icon