Picture for Ruizhe Chen

Ruizhe Chen

FairSteer: Inference Time Debiasing for LLMs with Dynamic Activation Steering

Add code
Apr 20, 2025
Viaarxiv icon

Persona-judge: Personalized Alignment of Large Language Models via Token-level Self-judgment

Add code
Apr 17, 2025
Viaarxiv icon

An All-Atom Generative Model for Designing Protein Complexes

Add code
Apr 17, 2025
Viaarxiv icon

MT-R1-Zero: Advancing LLM-based Machine Translation via R1-Zero-like Reinforcement Learning

Add code
Apr 14, 2025
Viaarxiv icon

DiffPO: Diffusion-styled Preference Optimization for Efficient Inference-Time Alignment of Large Language Models

Add code
Mar 06, 2025
Viaarxiv icon

Iterative Value Function Optimization for Guided Decoding

Add code
Mar 05, 2025
Viaarxiv icon

Performance Evaluation of Large Language Models in Statistical Programming

Add code
Feb 18, 2025
Viaarxiv icon

Decoupling Dark Knowledge via Block-wise Logit Distillation for Feature-level Alignment

Add code
Nov 03, 2024
Figure 1 for Decoupling Dark Knowledge via Block-wise Logit Distillation for Feature-level Alignment
Figure 2 for Decoupling Dark Knowledge via Block-wise Logit Distillation for Feature-level Alignment
Figure 3 for Decoupling Dark Knowledge via Block-wise Logit Distillation for Feature-level Alignment
Figure 4 for Decoupling Dark Knowledge via Block-wise Logit Distillation for Feature-level Alignment
Viaarxiv icon

FairMT-Bench: Benchmarking Fairness for Multi-turn Dialogue in Conversational LLMs

Add code
Oct 25, 2024
Figure 1 for FairMT-Bench: Benchmarking Fairness for Multi-turn Dialogue in Conversational LLMs
Figure 2 for FairMT-Bench: Benchmarking Fairness for Multi-turn Dialogue in Conversational LLMs
Figure 3 for FairMT-Bench: Benchmarking Fairness for Multi-turn Dialogue in Conversational LLMs
Figure 4 for FairMT-Bench: Benchmarking Fairness for Multi-turn Dialogue in Conversational LLMs
Viaarxiv icon

Modality-Fair Preference Optimization for Trustworthy MLLM Alignment

Add code
Oct 20, 2024
Viaarxiv icon