Picture for Akifumi Wachi

Akifumi Wachi

A Provable Approach for End-to-End Safe Reinforcement Learning

Add code
May 28, 2025
Viaarxiv icon

Offline Guarded Safe Reinforcement Learning for Medical Treatment Optimization Strategies

Add code
May 22, 2025
Viaarxiv icon

Target Return Optimizer for Multi-Game Decision Transformer

Add code
Mar 04, 2025
Viaarxiv icon

Vulnerability Mitigation for Safety-Aligned Language Models via Debiasing

Add code
Feb 04, 2025
Viaarxiv icon

Flipping-based Policy for Chance-Constrained Markov Decision Processes

Add code
Oct 09, 2024
Viaarxiv icon

Stepwise Alignment for Constrained Language Model Policy Optimization

Add code
Apr 17, 2024
Viaarxiv icon

A Survey of Constraint Formulations in Safe Reinforcement Learning

Add code
Feb 03, 2024
Viaarxiv icon

Long-term Safe Reinforcement Learning with Binary Feedback

Add code
Jan 11, 2024
Viaarxiv icon

Verbosity Bias in Preference Labeling by Large Language Models

Add code
Oct 16, 2023
Figure 1 for Verbosity Bias in Preference Labeling by Large Language Models
Figure 2 for Verbosity Bias in Preference Labeling by Large Language Models
Figure 3 for Verbosity Bias in Preference Labeling by Large Language Models
Figure 4 for Verbosity Bias in Preference Labeling by Large Language Models
Viaarxiv icon

Safe Exploration in Reinforcement Learning: A Generalized Formulation and Algorithms

Add code
Oct 05, 2023
Viaarxiv icon