Picture for Xuchao Zhang

Xuchao Zhang

LEGOMem: Modular Procedural Memory for Multi-agent LLM Systems for Workflow Automation

Add code
Oct 06, 2025
Viaarxiv icon

Enhancing Reasoning Capabilities of Small Language Models with Blueprints and Prompt Template Search

Add code
Jun 10, 2025
Viaarxiv icon

Anyprefer: An Agentic Framework for Preference Data Synthesis

Add code
Apr 27, 2025
Figure 1 for Anyprefer: An Agentic Framework for Preference Data Synthesis
Figure 2 for Anyprefer: An Agentic Framework for Preference Data Synthesis
Figure 3 for Anyprefer: An Agentic Framework for Preference Data Synthesis
Figure 4 for Anyprefer: An Agentic Framework for Preference Data Synthesis
Viaarxiv icon

Synergistic Weak-Strong Collaboration by Aligning Preferences

Add code
Apr 22, 2025
Viaarxiv icon

AMPO: Active Multi-Preference Optimization

Add code
Feb 25, 2025
Viaarxiv icon

Verifiable Format Control for Large Language Model Generations

Add code
Feb 06, 2025
Viaarxiv icon

REFA: Reference Free Alignment for multi-preference optimization

Add code
Dec 20, 2024
Viaarxiv icon

SWEPO: Simultaneous Weighted Preference Optimization for Group Contrastive Alignment

Add code
Dec 05, 2024
Viaarxiv icon

Unveiling Context-Aware Criteria in Self-Assessing LLMs

Add code
Oct 28, 2024
Figure 1 for Unveiling Context-Aware Criteria in Self-Assessing LLMs
Figure 2 for Unveiling Context-Aware Criteria in Self-Assessing LLMs
Figure 3 for Unveiling Context-Aware Criteria in Self-Assessing LLMs
Figure 4 for Unveiling Context-Aware Criteria in Self-Assessing LLMs
Viaarxiv icon

CREAM: Consistency Regularized Self-Rewarding Language Models

Add code
Oct 17, 2024
Figure 1 for CREAM: Consistency Regularized Self-Rewarding Language Models
Figure 2 for CREAM: Consistency Regularized Self-Rewarding Language Models
Figure 3 for CREAM: Consistency Regularized Self-Rewarding Language Models
Figure 4 for CREAM: Consistency Regularized Self-Rewarding Language Models
Viaarxiv icon