Alert button
Picture for Eiji Uchibe

Eiji Uchibe

Alert button

Randomized-to-Canonical Model Predictive Control for Real-world Visual Robotic Manipulation

Add code
Bookmark button
Alert button
Jul 05, 2022
Tomoya Yamanokuchi, Yuhwan Kwon, Yoshihisa Tsurumine, Eiji Uchibe, Jun Morimoto, Takamitsu Matsubara

Figure 1 for Randomized-to-Canonical Model Predictive Control for Real-world Visual Robotic Manipulation
Figure 2 for Randomized-to-Canonical Model Predictive Control for Real-world Visual Robotic Manipulation
Figure 3 for Randomized-to-Canonical Model Predictive Control for Real-world Visual Robotic Manipulation
Figure 4 for Randomized-to-Canonical Model Predictive Control for Real-world Visual Robotic Manipulation
Viaarxiv icon

Model-Based Imitation Learning Using Entropy Regularization of Model and Policy

Add code
Bookmark button
Alert button
Jun 21, 2022
Eiji Uchibe

Figure 1 for Model-Based Imitation Learning Using Entropy Regularization of Model and Policy
Figure 2 for Model-Based Imitation Learning Using Entropy Regularization of Model and Policy
Figure 3 for Model-Based Imitation Learning Using Entropy Regularization of Model and Policy
Figure 4 for Model-Based Imitation Learning Using Entropy Regularization of Model and Policy
Viaarxiv icon

$q$-Munchausen Reinforcement Learning

Add code
Bookmark button
Alert button
May 16, 2022
Lingwei Zhu, Zheng Chen, Eiji Uchibe, Takamitsu Matsubara

Figure 1 for $q$-Munchausen Reinforcement Learning
Figure 2 for $q$-Munchausen Reinforcement Learning
Figure 3 for $q$-Munchausen Reinforcement Learning
Figure 4 for $q$-Munchausen Reinforcement Learning
Viaarxiv icon

Enforcing KL Regularization in General Tsallis Entropy Reinforcement Learning via Advantage Learning

Add code
Bookmark button
Alert button
May 16, 2022
Lingwei Zhu, Zheng Chen, Eiji Uchibe, Takamitsu Matsubara

Figure 1 for Enforcing KL Regularization in General Tsallis Entropy Reinforcement Learning via Advantage Learning
Figure 2 for Enforcing KL Regularization in General Tsallis Entropy Reinforcement Learning via Advantage Learning
Figure 3 for Enforcing KL Regularization in General Tsallis Entropy Reinforcement Learning via Advantage Learning
Figure 4 for Enforcing KL Regularization in General Tsallis Entropy Reinforcement Learning via Advantage Learning
Viaarxiv icon

Imitation learning based on entropy-regularized forward and inverse reinforcement learning

Add code
Bookmark button
Alert button
Aug 17, 2020
Eiji Uchibe, Kenji Doya

Figure 1 for Imitation learning based on entropy-regularized forward and inverse reinforcement learning
Figure 2 for Imitation learning based on entropy-regularized forward and inverse reinforcement learning
Figure 3 for Imitation learning based on entropy-regularized forward and inverse reinforcement learning
Figure 4 for Imitation learning based on entropy-regularized forward and inverse reinforcement learning
Viaarxiv icon

Unbounded Output Networks for Classification

Add code
Bookmark button
Alert button
Jul 25, 2018
Stefan Elfwing, Eiji Uchibe, Kenji Doya

Figure 1 for Unbounded Output Networks for Classification
Figure 2 for Unbounded Output Networks for Classification
Figure 3 for Unbounded Output Networks for Classification
Figure 4 for Unbounded Output Networks for Classification
Viaarxiv icon

Unifying Value Iteration, Advantage Learning, and Dynamic Policy Programming

Add code
Bookmark button
Alert button
Oct 30, 2017
Tadashi Kozuno, Eiji Uchibe, Kenji Doya

Figure 1 for Unifying Value Iteration, Advantage Learning, and Dynamic Policy Programming
Figure 2 for Unifying Value Iteration, Advantage Learning, and Dynamic Policy Programming
Figure 3 for Unifying Value Iteration, Advantage Learning, and Dynamic Policy Programming
Figure 4 for Unifying Value Iteration, Advantage Learning, and Dynamic Policy Programming
Viaarxiv icon

Online Meta-learning by Parallel Algorithm Competition

Add code
Bookmark button
Alert button
Feb 24, 2017
Stefan Elfwing, Eiji Uchibe, Kenji Doya

Figure 1 for Online Meta-learning by Parallel Algorithm Competition
Figure 2 for Online Meta-learning by Parallel Algorithm Competition
Figure 3 for Online Meta-learning by Parallel Algorithm Competition
Figure 4 for Online Meta-learning by Parallel Algorithm Competition
Viaarxiv icon