Alert button
Picture for Haruka Kiyohara

Haruka Kiyohara

Alert button

Off-Policy Evaluation of Slate Bandit Policies via Optimizing Abstraction

Add code
Bookmark button
Alert button
Feb 03, 2024
Haruka Kiyohara, Masahiro Nomura, Yuta Saito

Viaarxiv icon

SCOPE-RL: A Python Library for Offline Reinforcement Learning and Off-Policy Evaluation

Add code
Bookmark button
Alert button
Dec 04, 2023
Haruka Kiyohara, Ren Kishimoto, Kosuke Kawakami, Ken Kobayashi, Kazuhide Nakata, Yuta Saito

Figure 1 for SCOPE-RL: A Python Library for Offline Reinforcement Learning and Off-Policy Evaluation
Figure 2 for SCOPE-RL: A Python Library for Offline Reinforcement Learning and Off-Policy Evaluation
Figure 3 for SCOPE-RL: A Python Library for Offline Reinforcement Learning and Off-Policy Evaluation
Figure 4 for SCOPE-RL: A Python Library for Offline Reinforcement Learning and Off-Policy Evaluation
Viaarxiv icon

Towards Assessing and Benchmarking Risk-Return Tradeoff of Off-Policy Evaluation

Add code
Bookmark button
Alert button
Dec 04, 2023
Haruka Kiyohara, Ren Kishimoto, Kosuke Kawakami, Ken Kobayashi, Kazuhide Nakata, Yuta Saito

Viaarxiv icon

Off-Policy Evaluation of Ranking Policies under Diverse User Behavior

Add code
Bookmark button
Alert button
Jun 26, 2023
Haruka Kiyohara, Masatoshi Uehara, Yusuke Narita, Nobuyuki Shimizu, Yasuo Yamamoto, Yuta Saito

Figure 1 for Off-Policy Evaluation of Ranking Policies under Diverse User Behavior
Figure 2 for Off-Policy Evaluation of Ranking Policies under Diverse User Behavior
Figure 3 for Off-Policy Evaluation of Ranking Policies under Diverse User Behavior
Figure 4 for Off-Policy Evaluation of Ranking Policies under Diverse User Behavior
Viaarxiv icon

Policy-Adaptive Estimator Selection for Off-Policy Evaluation

Add code
Bookmark button
Alert button
Nov 25, 2022
Takuma Udagawa, Haruka Kiyohara, Yusuke Narita, Yuta Saito, Kei Tateno

Figure 1 for Policy-Adaptive Estimator Selection for Off-Policy Evaluation
Figure 2 for Policy-Adaptive Estimator Selection for Off-Policy Evaluation
Figure 3 for Policy-Adaptive Estimator Selection for Off-Policy Evaluation
Figure 4 for Policy-Adaptive Estimator Selection for Off-Policy Evaluation
Viaarxiv icon

Future-Dependent Value-Based Off-Policy Evaluation in POMDPs

Add code
Bookmark button
Alert button
Jul 26, 2022
Masatoshi Uehara, Haruka Kiyohara, Andrew Bennett, Victor Chernozhukov, Nan Jiang, Nathan Kallus, Chengchun Shi, Wen Sun

Figure 1 for Future-Dependent Value-Based Off-Policy Evaluation in POMDPs
Figure 2 for Future-Dependent Value-Based Off-Policy Evaluation in POMDPs
Figure 3 for Future-Dependent Value-Based Off-Policy Evaluation in POMDPs
Figure 4 for Future-Dependent Value-Based Off-Policy Evaluation in POMDPs
Viaarxiv icon

Doubly Robust Off-Policy Evaluation for Ranking Policies under the Cascade Behavior Model

Add code
Bookmark button
Alert button
Feb 03, 2022
Haruka Kiyohara, Yuta Saito, Tatsuya Matsuhiro, Yusuke Narita, Nobuyuki Shimizu, Yasuo Yamamoto

Figure 1 for Doubly Robust Off-Policy Evaluation for Ranking Policies under the Cascade Behavior Model
Figure 2 for Doubly Robust Off-Policy Evaluation for Ranking Policies under the Cascade Behavior Model
Figure 3 for Doubly Robust Off-Policy Evaluation for Ranking Policies under the Cascade Behavior Model
Figure 4 for Doubly Robust Off-Policy Evaluation for Ranking Policies under the Cascade Behavior Model
Viaarxiv icon

Accelerating Offline Reinforcement Learning Application in Real-Time Bidding and Recommendation: Potential Use of Simulation

Add code
Bookmark button
Alert button
Sep 17, 2021
Haruka Kiyohara, Kosuke Kawakami, Yuta Saito

Figure 1 for Accelerating Offline Reinforcement Learning Application in Real-Time Bidding and Recommendation: Potential Use of Simulation
Figure 2 for Accelerating Offline Reinforcement Learning Application in Real-Time Bidding and Recommendation: Potential Use of Simulation
Viaarxiv icon