Picture for Zhanhui Zhou

Zhanhui Zhou

SafeCoT: Improving VLM Safety with Minimal Reasoning

Add code
Jun 11, 2025
Viaarxiv icon

RePO: Replay-Enhanced Policy Optimization

Add code
Jun 11, 2025
Viaarxiv icon

Mitigating Object Hallucination via Robust Local Perception Search

Add code
Jun 07, 2025
Viaarxiv icon

Emergent Response Planning in LLM

Add code
Feb 10, 2025
Figure 1 for Emergent Response Planning in LLM
Figure 2 for Emergent Response Planning in LLM
Figure 3 for Emergent Response Planning in LLM
Figure 4 for Emergent Response Planning in LLM
Viaarxiv icon

Inference-Time Language Model Alignment via Integrated Value Guidance

Add code
Sep 26, 2024
Viaarxiv icon

Iterative Length-Regularized Direct Preference Optimization: A Case Study on Improving 7B Language Models to GPT-4 Level

Add code
Jun 17, 2024
Viaarxiv icon

Weak-to-Strong Search: Align Large Language Models via Searching over Small Language Models

Add code
May 29, 2024
Figure 1 for Weak-to-Strong Search: Align Large Language Models via Searching over Small Language Models
Figure 2 for Weak-to-Strong Search: Align Large Language Models via Searching over Small Language Models
Figure 3 for Weak-to-Strong Search: Align Large Language Models via Searching over Small Language Models
Figure 4 for Weak-to-Strong Search: Align Large Language Models via Searching over Small Language Models
Viaarxiv icon

ConceptMath: A Bilingual Concept-wise Benchmark for Measuring Mathematical Reasoning of Large Language Models

Add code
Feb 23, 2024
Figure 1 for ConceptMath: A Bilingual Concept-wise Benchmark for Measuring Mathematical Reasoning of Large Language Models
Figure 2 for ConceptMath: A Bilingual Concept-wise Benchmark for Measuring Mathematical Reasoning of Large Language Models
Figure 3 for ConceptMath: A Bilingual Concept-wise Benchmark for Measuring Mathematical Reasoning of Large Language Models
Figure 4 for ConceptMath: A Bilingual Concept-wise Benchmark for Measuring Mathematical Reasoning of Large Language Models
Viaarxiv icon

MT-Bench-101: A Fine-Grained Benchmark for Evaluating Large Language Models in Multi-Turn Dialogues

Add code
Feb 22, 2024
Figure 1 for MT-Bench-101: A Fine-Grained Benchmark for Evaluating Large Language Models in Multi-Turn Dialogues
Figure 2 for MT-Bench-101: A Fine-Grained Benchmark for Evaluating Large Language Models in Multi-Turn Dialogues
Figure 3 for MT-Bench-101: A Fine-Grained Benchmark for Evaluating Large Language Models in Multi-Turn Dialogues
Figure 4 for MT-Bench-101: A Fine-Grained Benchmark for Evaluating Large Language Models in Multi-Turn Dialogues
Viaarxiv icon

Emulated Disalignment: Safety Alignment for Large Language Models May Backfire!

Add code
Feb 21, 2024
Figure 1 for Emulated Disalignment: Safety Alignment for Large Language Models May Backfire!
Figure 2 for Emulated Disalignment: Safety Alignment for Large Language Models May Backfire!
Figure 3 for Emulated Disalignment: Safety Alignment for Large Language Models May Backfire!
Figure 4 for Emulated Disalignment: Safety Alignment for Large Language Models May Backfire!
Viaarxiv icon