Picture for Yaodong Yang

Yaodong Yang

Aligner: Achieving Efficient Alignment through Weak-to-Strong Correction

Add code
Feb 06, 2024
Viaarxiv icon

Panacea: Pareto Alignment via Preference Adaptation for LLMs

Add code
Feb 03, 2024
Figure 1 for Panacea: Pareto Alignment via Preference Adaptation for LLMs
Figure 2 for Panacea: Pareto Alignment via Preference Adaptation for LLMs
Figure 3 for Panacea: Pareto Alignment via Preference Adaptation for LLMs
Figure 4 for Panacea: Pareto Alignment via Preference Adaptation for LLMs
Viaarxiv icon

CivRealm: A Learning and Reasoning Odyssey in Civilization for Decision-Making Agents

Add code
Jan 19, 2024
Figure 1 for CivRealm: A Learning and Reasoning Odyssey in Civilization for Decision-Making Agents
Figure 2 for CivRealm: A Learning and Reasoning Odyssey in Civilization for Decision-Making Agents
Figure 3 for CivRealm: A Learning and Reasoning Odyssey in Civilization for Decision-Making Agents
Figure 4 for CivRealm: A Learning and Reasoning Odyssey in Civilization for Decision-Making Agents
Viaarxiv icon

A Perspective of Q-value Estimation on Offline-to-Online Reinforcement Learning

Add code
Dec 12, 2023
Figure 1 for A Perspective of Q-value Estimation on Offline-to-Online Reinforcement Learning
Figure 2 for A Perspective of Q-value Estimation on Offline-to-Online Reinforcement Learning
Figure 3 for A Perspective of Q-value Estimation on Offline-to-Online Reinforcement Learning
Figure 4 for A Perspective of Q-value Estimation on Offline-to-Online Reinforcement Learning
Viaarxiv icon

JARVIS-1: Open-World Multi-task Agents with Memory-Augmented Multimodal Language Models

Add code
Nov 30, 2023
Viaarxiv icon

AI Alignment: A Comprehensive Survey

Add code
Nov 01, 2023
Viaarxiv icon

Grasp Multiple Objects with One Hand

Add code
Oct 24, 2023
Viaarxiv icon

Measuring Value Understanding in Language Models through Discriminator-Critique Gap

Add code
Oct 19, 2023
Viaarxiv icon

Safety-Gymnasium: A Unified Safe Reinforcement Learning Benchmark

Add code
Oct 19, 2023
Figure 1 for Safety-Gymnasium: A Unified Safe Reinforcement Learning Benchmark
Figure 2 for Safety-Gymnasium: A Unified Safe Reinforcement Learning Benchmark
Figure 3 for Safety-Gymnasium: A Unified Safe Reinforcement Learning Benchmark
Figure 4 for Safety-Gymnasium: A Unified Safe Reinforcement Learning Benchmark
Viaarxiv icon

Safe RLHF: Safe Reinforcement Learning from Human Feedback

Add code
Oct 19, 2023
Figure 1 for Safe RLHF: Safe Reinforcement Learning from Human Feedback
Figure 2 for Safe RLHF: Safe Reinforcement Learning from Human Feedback
Figure 3 for Safe RLHF: Safe Reinforcement Learning from Human Feedback
Figure 4 for Safe RLHF: Safe Reinforcement Learning from Human Feedback
Viaarxiv icon