Picture for Wenhong Zhu

Wenhong Zhu

MrRoPE: Mixed-radix Rotary Position Embedding

Add code
Jan 28, 2026
Viaarxiv icon

InnovatorBench: Evaluating Agents' Ability to Conduct Innovative LLM Research

Add code
Nov 03, 2025
Figure 1 for InnovatorBench: Evaluating Agents' Ability to Conduct Innovative LLM Research
Figure 2 for InnovatorBench: Evaluating Agents' Ability to Conduct Innovative LLM Research
Figure 3 for InnovatorBench: Evaluating Agents' Ability to Conduct Innovative LLM Research
Figure 4 for InnovatorBench: Evaluating Agents' Ability to Conduct Innovative LLM Research
Viaarxiv icon

Proximal Supervised Fine-Tuning

Add code
Aug 25, 2025
Figure 1 for Proximal Supervised Fine-Tuning
Figure 2 for Proximal Supervised Fine-Tuning
Figure 3 for Proximal Supervised Fine-Tuning
Figure 4 for Proximal Supervised Fine-Tuning
Viaarxiv icon

Flexible Realignment of Language Models

Add code
Jun 15, 2025
Figure 1 for Flexible Realignment of Language Models
Figure 2 for Flexible Realignment of Language Models
Figure 3 for Flexible Realignment of Language Models
Figure 4 for Flexible Realignment of Language Models
Viaarxiv icon

Adding Alignment Control to Language Models

Add code
Mar 07, 2025
Figure 1 for Adding Alignment Control to Language Models
Figure 2 for Adding Alignment Control to Language Models
Figure 3 for Adding Alignment Control to Language Models
Figure 4 for Adding Alignment Control to Language Models
Viaarxiv icon

Do Large Language Models Truly Understand Geometric Structures?

Add code
Jan 23, 2025
Figure 1 for Do Large Language Models Truly Understand Geometric Structures?
Figure 2 for Do Large Language Models Truly Understand Geometric Structures?
Figure 3 for Do Large Language Models Truly Understand Geometric Structures?
Figure 4 for Do Large Language Models Truly Understand Geometric Structures?
Viaarxiv icon

Weak-to-Strong Preference Optimization: Stealing Reward from Weak Aligned Model

Add code
Oct 24, 2024
Figure 1 for Weak-to-Strong Preference Optimization: Stealing Reward from Weak Aligned Model
Figure 2 for Weak-to-Strong Preference Optimization: Stealing Reward from Weak Aligned Model
Figure 3 for Weak-to-Strong Preference Optimization: Stealing Reward from Weak Aligned Model
Figure 4 for Weak-to-Strong Preference Optimization: Stealing Reward from Weak Aligned Model
Viaarxiv icon

Improving Open-Ended Text Generation via Adaptive Decoding

Add code
Feb 28, 2024
Figure 1 for Improving Open-Ended Text Generation via Adaptive Decoding
Figure 2 for Improving Open-Ended Text Generation via Adaptive Decoding
Figure 3 for Improving Open-Ended Text Generation via Adaptive Decoding
Figure 4 for Improving Open-Ended Text Generation via Adaptive Decoding
Viaarxiv icon

Is Cognition and Action Consistent or Not: Investigating Large Language Model's Personality

Add code
Feb 22, 2024
Viaarxiv icon

CLEAN-EVAL: Clean Evaluation on Contaminated Large Language Models

Add code
Nov 15, 2023
Figure 1 for CLEAN-EVAL: Clean Evaluation on Contaminated Large Language Models
Figure 2 for CLEAN-EVAL: Clean Evaluation on Contaminated Large Language Models
Figure 3 for CLEAN-EVAL: Clean Evaluation on Contaminated Large Language Models
Figure 4 for CLEAN-EVAL: Clean Evaluation on Contaminated Large Language Models
Viaarxiv icon