Picture for David D. Yao

David D. Yao

Fine-Tuning Diffusion Generative Models via Rich Preference Optimization

Add code
Mar 13, 2025
Viaarxiv icon

Score as Action: Fine-Tuning Diffusion Generative Models by Continuous-time Reinforcement Learning

Add code
Feb 03, 2025
Figure 1 for Score as Action: Fine-Tuning Diffusion Generative Models by Continuous-time Reinforcement Learning
Figure 2 for Score as Action: Fine-Tuning Diffusion Generative Models by Continuous-time Reinforcement Learning
Figure 3 for Score as Action: Fine-Tuning Diffusion Generative Models by Continuous-time Reinforcement Learning
Figure 4 for Score as Action: Fine-Tuning Diffusion Generative Models by Continuous-time Reinforcement Learning
Viaarxiv icon

RainbowPO: A Unified Framework for Combining Improvements in Preference Optimization

Add code
Oct 05, 2024
Figure 1 for RainbowPO: A Unified Framework for Combining Improvements in Preference Optimization
Figure 2 for RainbowPO: A Unified Framework for Combining Improvements in Preference Optimization
Figure 3 for RainbowPO: A Unified Framework for Combining Improvements in Preference Optimization
Figure 4 for RainbowPO: A Unified Framework for Combining Improvements in Preference Optimization
Viaarxiv icon

Preference Tuning with Human Feedback on Language, Speech, and Vision Tasks: A Survey

Add code
Sep 17, 2024
Viaarxiv icon

Scores as Actions: a framework of fine-tuning diffusion models by continuous-time reinforcement learning

Add code
Sep 12, 2024
Viaarxiv icon

Policy Optimization for Continuous Reinforcement Learning

Add code
Jun 02, 2023
Figure 1 for Policy Optimization for Continuous Reinforcement Learning
Figure 2 for Policy Optimization for Continuous Reinforcement Learning
Figure 3 for Policy Optimization for Continuous Reinforcement Learning
Figure 4 for Policy Optimization for Continuous Reinforcement Learning
Viaarxiv icon