Picture for Shuchang Tao

Shuchang Tao

d-TreeRPO: Towards More Reliable Policy Optimization for Diffusion Language Models

Add code
Dec 10, 2025
Viaarxiv icon

AgentEvolver: Towards Efficient Self-Evolving Agent System

Add code
Nov 13, 2025
Viaarxiv icon

Omni-SafetyBench: A Benchmark for Safety Evaluation of Audio-Visual Large Language Models

Add code
Aug 10, 2025
Figure 1 for Omni-SafetyBench: A Benchmark for Safety Evaluation of Audio-Visual Large Language Models
Figure 2 for Omni-SafetyBench: A Benchmark for Safety Evaluation of Audio-Visual Large Language Models
Figure 3 for Omni-SafetyBench: A Benchmark for Safety Evaluation of Audio-Visual Large Language Models
Figure 4 for Omni-SafetyBench: A Benchmark for Safety Evaluation of Audio-Visual Large Language Models
Viaarxiv icon

Incentivizing Strong Reasoning from Weak Supervision

Add code
May 28, 2025
Figure 1 for Incentivizing Strong Reasoning from Weak Supervision
Figure 2 for Incentivizing Strong Reasoning from Weak Supervision
Figure 3 for Incentivizing Strong Reasoning from Weak Supervision
Figure 4 for Incentivizing Strong Reasoning from Weak Supervision
Viaarxiv icon

Inference-time Alignment in Continuous Space

Add code
May 26, 2025
Viaarxiv icon

Incentivizing Reasoning from Weak Supervision

Add code
May 26, 2025
Figure 1 for Incentivizing Reasoning from Weak Supervision
Figure 2 for Incentivizing Reasoning from Weak Supervision
Figure 3 for Incentivizing Reasoning from Weak Supervision
Figure 4 for Incentivizing Reasoning from Weak Supervision
Viaarxiv icon

ToolCoder: A Systematic Code-Empowered Tool Learning Framework for Large Language Models

Add code
Feb 17, 2025
Figure 1 for ToolCoder: A Systematic Code-Empowered Tool Learning Framework for Large Language Models
Figure 2 for ToolCoder: A Systematic Code-Empowered Tool Learning Framework for Large Language Models
Figure 3 for ToolCoder: A Systematic Code-Empowered Tool Learning Framework for Large Language Models
Figure 4 for ToolCoder: A Systematic Code-Empowered Tool Learning Framework for Large Language Models
Viaarxiv icon

Revisiting Robust RAG: Do We Still Need Complex Robust Training in the Era of Powerful LLMs?

Add code
Feb 17, 2025
Viaarxiv icon

Accelerating the Surrogate Retraining for Poisoning Attacks against Recommender Systems

Add code
Aug 20, 2024
Figure 1 for Accelerating the Surrogate Retraining for Poisoning Attacks against Recommender Systems
Figure 2 for Accelerating the Surrogate Retraining for Poisoning Attacks against Recommender Systems
Figure 3 for Accelerating the Surrogate Retraining for Poisoning Attacks against Recommender Systems
Figure 4 for Accelerating the Surrogate Retraining for Poisoning Attacks against Recommender Systems
Viaarxiv icon

When to Trust LLMs: Aligning Confidence with Response Quality

Add code
Apr 26, 2024
Viaarxiv icon