Picture for Boyuan Chen

Boyuan Chen

Jack

InterMT: Multi-Turn Interleaved Preference Alignment with Human Feedback

Add code
May 29, 2025
Viaarxiv icon

Generative RLHF-V: Learning Principles from Multi-modal Human Preference

Add code
May 24, 2025
Viaarxiv icon

Mitigating Deceptive Alignment via Self-Monitoring

Add code
May 24, 2025
Viaarxiv icon

LAPP: Large Language Model Feedback for Preference-Driven Reinforcement Learning

Add code
Apr 21, 2025
Viaarxiv icon

Empirical Analysis of Sim-and-Real Cotraining Of Diffusion Policies For Planar Pushing from Pixels

Add code
Mar 28, 2025
Viaarxiv icon

PhysGen3D: Crafting a Miniature Interactive World from a Single Image

Add code
Mar 26, 2025
Viaarxiv icon

History-Guided Video Diffusion

Add code
Feb 10, 2025
Viaarxiv icon

RedStar: Does Scaling Long-CoT Data Unlock Better Slow-Reasoning Systems?

Add code
Jan 20, 2025
Viaarxiv icon

A Paragraph is All It Takes: Rich Robot Behaviors from Interacting, Trusted LLMs

Add code
Dec 24, 2024
Viaarxiv icon

Align Anything: Training All-Modality Models to Follow Instructions with Language Feedback

Add code
Dec 20, 2024
Figure 1 for Align Anything: Training All-Modality Models to Follow Instructions with Language Feedback
Figure 2 for Align Anything: Training All-Modality Models to Follow Instructions with Language Feedback
Figure 3 for Align Anything: Training All-Modality Models to Follow Instructions with Language Feedback
Figure 4 for Align Anything: Training All-Modality Models to Follow Instructions with Language Feedback
Viaarxiv icon