Picture for Yaodong Yang

Yaodong Yang

Mitigating Deceptive Alignment via Self-Monitoring

Add code
May 24, 2025
Viaarxiv icon

Generative RLHF-V: Learning Principles from Multi-modal Human Preference

Add code
May 24, 2025
Viaarxiv icon

Adaptive Visuo-Tactile Fusion with Predictive Force Attention for Dexterous Manipulation

Add code
May 20, 2025
Viaarxiv icon

J1: Exploring Simple Test-Time Scaling for LLM-as-a-Judge

Add code
May 17, 2025
Figure 1 for J1: Exploring Simple Test-Time Scaling for LLM-as-a-Judge
Figure 2 for J1: Exploring Simple Test-Time Scaling for LLM-as-a-Judge
Figure 3 for J1: Exploring Simple Test-Time Scaling for LLM-as-a-Judge
Figure 4 for J1: Exploring Simple Test-Time Scaling for LLM-as-a-Judge
Viaarxiv icon

Redefining Superalignment: From Weak-to-Strong Alignment to Human-AI Co-Alignment to Sustainable Symbiotic Society

Add code
Apr 24, 2025
Viaarxiv icon

A Comprehensive Survey in LLM(-Agent) Full Stack Safety: Data, Training and Deployment

Add code
Apr 22, 2025
Viaarxiv icon

Benchmarking Multi-National Value Alignment for Large Language Models

Add code
Apr 19, 2025
Figure 1 for Benchmarking Multi-National Value Alignment for Large Language Models
Figure 2 for Benchmarking Multi-National Value Alignment for Large Language Models
Figure 3 for Benchmarking Multi-National Value Alignment for Large Language Models
Figure 4 for Benchmarking Multi-National Value Alignment for Large Language Models
Viaarxiv icon

Dexterous Non-Prehensile Manipulation for Ungraspable Object via Extrinsic Dexterity

Add code
Mar 29, 2025
Figure 1 for Dexterous Non-Prehensile Manipulation for Ungraspable Object via Extrinsic Dexterity
Figure 2 for Dexterous Non-Prehensile Manipulation for Ungraspable Object via Extrinsic Dexterity
Figure 3 for Dexterous Non-Prehensile Manipulation for Ungraspable Object via Extrinsic Dexterity
Figure 4 for Dexterous Non-Prehensile Manipulation for Ungraspable Object via Extrinsic Dexterity
Viaarxiv icon

ThinkPatterns-21k: A Systematic Study on the Impact of Thinking Patterns in LLMs

Add code
Mar 17, 2025
Viaarxiv icon

SafeVLA: Towards Safety Alignment of Vision-Language-Action Model via Safe Reinforcement Learning

Add code
Mar 05, 2025
Figure 1 for SafeVLA: Towards Safety Alignment of Vision-Language-Action Model via Safe Reinforcement Learning
Figure 2 for SafeVLA: Towards Safety Alignment of Vision-Language-Action Model via Safe Reinforcement Learning
Figure 3 for SafeVLA: Towards Safety Alignment of Vision-Language-Action Model via Safe Reinforcement Learning
Figure 4 for SafeVLA: Towards Safety Alignment of Vision-Language-Action Model via Safe Reinforcement Learning
Viaarxiv icon