Picture for Zhuo Li

Zhuo Li

MARCH: Multi-Agent Reinforced Self-Check for LLM Hallucination

Add code
Mar 25, 2026
Viaarxiv icon

Towards Cold-Start Drafting and Continual Refining: A Value-Driven Memory Approach with Application to NPU Kernel Synthesis

Add code
Mar 11, 2026
Viaarxiv icon

Mitigating Reward Hacking in RLHF via Bayesian Non-negative Reward Modeling

Add code
Feb 11, 2026
Viaarxiv icon

Echoes as Anchors: Probabilistic Costs and Attention Refocusing in LLM Reasoning

Add code
Feb 06, 2026
Viaarxiv icon

Dual-Prototype Disentanglement: A Context-Aware Enhancement Framework for Time Series Forecasting

Add code
Jan 27, 2026
Viaarxiv icon

Safeguarding LLM Fine-tuning via Push-Pull Distributional Alignment

Add code
Jan 12, 2026
Viaarxiv icon

Eliminating Inductive Bias in Reward Models with Information-Theoretic Guidance

Add code
Dec 29, 2025
Viaarxiv icon

Towards Deploying VLA without Fine-Tuning: Plug-and-Play Inference-Time VLA Policy Steering via Embodied Evolutionary Diffusion

Add code
Nov 18, 2025
Viaarxiv icon

The 'Sure' Trap: Multi-Scale Poisoning Analysis of Stealthy Compliance-Only Backdoors in Fine-Tuned Large Language Models

Add code
Nov 16, 2025
Figure 1 for The 'Sure' Trap: Multi-Scale Poisoning Analysis of Stealthy Compliance-Only Backdoors in Fine-Tuned Large Language Models
Figure 2 for The 'Sure' Trap: Multi-Scale Poisoning Analysis of Stealthy Compliance-Only Backdoors in Fine-Tuned Large Language Models
Figure 3 for The 'Sure' Trap: Multi-Scale Poisoning Analysis of Stealthy Compliance-Only Backdoors in Fine-Tuned Large Language Models
Figure 4 for The 'Sure' Trap: Multi-Scale Poisoning Analysis of Stealthy Compliance-Only Backdoors in Fine-Tuned Large Language Models
Viaarxiv icon

RECAP: Reproducing Copyrighted Data from LLMs Training with an Agentic Pipeline

Add code
Oct 29, 2025
Viaarxiv icon