Picture for Chan-Hung Yu

Chan-Hung Yu

Adaptive Helpfulness-Harmlessness Alignment with Preference Vectors

Add code
Apr 27, 2025
Figure 1 for Adaptive Helpfulness-Harmlessness Alignment with Preference Vectors
Figure 2 for Adaptive Helpfulness-Harmlessness Alignment with Preference Vectors
Figure 3 for Adaptive Helpfulness-Harmlessness Alignment with Preference Vectors
Figure 4 for Adaptive Helpfulness-Harmlessness Alignment with Preference Vectors
Viaarxiv icon

Synthesizing Programmatic Reinforcement Learning Policies with Large Language Model Guided Search

Add code
May 26, 2024
Figure 1 for Synthesizing Programmatic Reinforcement Learning Policies with Large Language Model Guided Search
Figure 2 for Synthesizing Programmatic Reinforcement Learning Policies with Large Language Model Guided Search
Figure 3 for Synthesizing Programmatic Reinforcement Learning Policies with Large Language Model Guided Search
Figure 4 for Synthesizing Programmatic Reinforcement Learning Policies with Large Language Model Guided Search
Viaarxiv icon

LLM Discussion: Enhancing the Creativity of Large Language Models via Discussion Framework and Role-Play

Add code
May 10, 2024
Figure 1 for LLM Discussion: Enhancing the Creativity of Large Language Models via Discussion Framework and Role-Play
Figure 2 for LLM Discussion: Enhancing the Creativity of Large Language Models via Discussion Framework and Role-Play
Figure 3 for LLM Discussion: Enhancing the Creativity of Large Language Models via Discussion Framework and Role-Play
Figure 4 for LLM Discussion: Enhancing the Creativity of Large Language Models via Discussion Framework and Role-Play
Viaarxiv icon