Picture for Zhepeng Cen

Zhepeng Cen

Pushing Forward Pareto Frontiers of Proactive Agents with Behavioral Agentic Optimization

Add code
Feb 11, 2026
Viaarxiv icon

LoCoBench-Agent: An Interactive Benchmark for LLM Agents in Long-Context Software Engineering

Add code
Nov 17, 2025
Figure 1 for LoCoBench-Agent: An Interactive Benchmark for LLM Agents in Long-Context Software Engineering
Figure 2 for LoCoBench-Agent: An Interactive Benchmark for LLM Agents in Long-Context Software Engineering
Figure 3 for LoCoBench-Agent: An Interactive Benchmark for LLM Agents in Long-Context Software Engineering
Figure 4 for LoCoBench-Agent: An Interactive Benchmark for LLM Agents in Long-Context Software Engineering
Viaarxiv icon

LoCoBench: A Benchmark for Long-Context Large Language Models in Complex Software Engineering

Add code
Sep 11, 2025
Figure 1 for LoCoBench: A Benchmark for Long-Context Large Language Models in Complex Software Engineering
Figure 2 for LoCoBench: A Benchmark for Long-Context Large Language Models in Complex Software Engineering
Figure 3 for LoCoBench: A Benchmark for Long-Context Large Language Models in Complex Software Engineering
Figure 4 for LoCoBench: A Benchmark for Long-Context Large Language Models in Complex Software Engineering
Viaarxiv icon

Behavior Injection: Preparing Language Models for Reinforcement Learning

Add code
May 25, 2025
Viaarxiv icon

Signal, Image, or Symbolic: Exploring the Best Input Representation for Electrocardiogram-Language Models Through a Unified Framework

Add code
May 24, 2025
Viaarxiv icon

Safety is Not Only About Refusal: Reasoning-Enhanced Fine-tuning for Interpretable LLM Safety

Add code
Mar 06, 2025
Figure 1 for Safety is Not Only About Refusal: Reasoning-Enhanced Fine-tuning for Interpretable LLM Safety
Figure 2 for Safety is Not Only About Refusal: Reasoning-Enhanced Fine-tuning for Interpretable LLM Safety
Figure 3 for Safety is Not Only About Refusal: Reasoning-Enhanced Fine-tuning for Interpretable LLM Safety
Figure 4 for Safety is Not Only About Refusal: Reasoning-Enhanced Fine-tuning for Interpretable LLM Safety
Viaarxiv icon

Your Language Model May Think Too Rigidly: Achieving Reasoning Consistency with Symmetry-Enhanced Training

Add code
Feb 25, 2025
Figure 1 for Your Language Model May Think Too Rigidly: Achieving Reasoning Consistency with Symmetry-Enhanced Training
Figure 2 for Your Language Model May Think Too Rigidly: Achieving Reasoning Consistency with Symmetry-Enhanced Training
Figure 3 for Your Language Model May Think Too Rigidly: Achieving Reasoning Consistency with Symmetry-Enhanced Training
Figure 4 for Your Language Model May Think Too Rigidly: Achieving Reasoning Consistency with Symmetry-Enhanced Training
Viaarxiv icon

Bridging the Training-Inference Gap in LLMs by Leveraging Self-Generated Tokens

Add code
Oct 18, 2024
Figure 1 for Bridging the Training-Inference Gap in LLMs by Leveraging Self-Generated Tokens
Figure 2 for Bridging the Training-Inference Gap in LLMs by Leveraging Self-Generated Tokens
Figure 3 for Bridging the Training-Inference Gap in LLMs by Leveraging Self-Generated Tokens
Figure 4 for Bridging the Training-Inference Gap in LLMs by Leveraging Self-Generated Tokens
Viaarxiv icon

OASIS: Conditional Distribution Shaping for Offline Safe Reinforcement Learning

Add code
Jul 19, 2024
Figure 1 for OASIS: Conditional Distribution Shaping for Offline Safe Reinforcement Learning
Figure 2 for OASIS: Conditional Distribution Shaping for Offline Safe Reinforcement Learning
Figure 3 for OASIS: Conditional Distribution Shaping for Offline Safe Reinforcement Learning
Figure 4 for OASIS: Conditional Distribution Shaping for Offline Safe Reinforcement Learning
Viaarxiv icon

Feasibility Consistent Representation Learning for Safe Reinforcement Learning

Add code
May 20, 2024
Viaarxiv icon