Picture for Xiaoyuan Yi

Xiaoyuan Yi

Microsoft Research Asia

Does LLM Alignment Really Need Diversity? An Empirical Study of Adapting RLVR Methods for Moral Reasoning

Add code
Mar 11, 2026
Viaarxiv icon

Contextualized Privacy Defense for LLM Agents

Add code
Mar 03, 2026
Viaarxiv icon

On the Dynamics of Multi-Agent LLM Communities Driven by Value Diversity

Add code
Dec 11, 2025
Viaarxiv icon

NAMeGEn: Creative Name Generation via A Novel Agent-based Multiple Personalized Goal Enhancement Framework

Add code
Nov 19, 2025
Viaarxiv icon

Legal$Δ$: Enhancing Legal Reasoning in LLMs via Reinforcement Learning with Chain-of-Thought Guided Information Gain

Add code
Aug 17, 2025
Figure 1 for Legal$Δ$: Enhancing Legal Reasoning in LLMs via Reinforcement Learning with Chain-of-Thought Guided Information Gain
Figure 2 for Legal$Δ$: Enhancing Legal Reasoning in LLMs via Reinforcement Learning with Chain-of-Thought Guided Information Gain
Figure 3 for Legal$Δ$: Enhancing Legal Reasoning in LLMs via Reinforcement Learning with Chain-of-Thought Guided Information Gain
Figure 4 for Legal$Δ$: Enhancing Legal Reasoning in LLMs via Reinforcement Learning with Chain-of-Thought Guided Information Gain
Viaarxiv icon

IROTE: Human-like Traits Elicitation of Large Language Model via In-Context Self-Reflective Optimization

Add code
Aug 12, 2025
Figure 1 for IROTE: Human-like Traits Elicitation of Large Language Model via In-Context Self-Reflective Optimization
Figure 2 for IROTE: Human-like Traits Elicitation of Large Language Model via In-Context Self-Reflective Optimization
Figure 3 for IROTE: Human-like Traits Elicitation of Large Language Model via In-Context Self-Reflective Optimization
Figure 4 for IROTE: Human-like Traits Elicitation of Large Language Model via In-Context Self-Reflective Optimization
Viaarxiv icon

The Incomplete Bridge: How AI Research (Mis)Engages with Psychology

Add code
Jul 30, 2025
Viaarxiv icon

PICACO: Pluralistic In-Context Value Alignment of LLMs via Total Correlation Optimization

Add code
Jul 22, 2025
Figure 1 for PICACO: Pluralistic In-Context Value Alignment of LLMs via Total Correlation Optimization
Figure 2 for PICACO: Pluralistic In-Context Value Alignment of LLMs via Total Correlation Optimization
Figure 3 for PICACO: Pluralistic In-Context Value Alignment of LLMs via Total Correlation Optimization
Figure 4 for PICACO: Pluralistic In-Context Value Alignment of LLMs via Total Correlation Optimization
Viaarxiv icon

MotiveBench: How Far Are We From Human-Like Motivational Reasoning in Large Language Models?

Add code
Jun 16, 2025
Figure 1 for MotiveBench: How Far Are We From Human-Like Motivational Reasoning in Large Language Models?
Figure 2 for MotiveBench: How Far Are We From Human-Like Motivational Reasoning in Large Language Models?
Figure 3 for MotiveBench: How Far Are We From Human-Like Motivational Reasoning in Large Language Models?
Figure 4 for MotiveBench: How Far Are We From Human-Like Motivational Reasoning in Large Language Models?
Viaarxiv icon

Unintended Harms of Value-Aligned LLMs: Psychological and Empirical Insights

Add code
Jun 06, 2025
Figure 1 for Unintended Harms of Value-Aligned LLMs: Psychological and Empirical Insights
Figure 2 for Unintended Harms of Value-Aligned LLMs: Psychological and Empirical Insights
Figure 3 for Unintended Harms of Value-Aligned LLMs: Psychological and Empirical Insights
Figure 4 for Unintended Harms of Value-Aligned LLMs: Psychological and Empirical Insights
Viaarxiv icon