Picture for Yanyan Zhao

Yanyan Zhao

Large Language Model Agents Are Not Always Faithful Self-Evolvers

Add code
Jan 30, 2026
Viaarxiv icon

TEA-Bench: A Systematic Benchmarking of Tool-enhanced Emotional Support Dialogue Agent

Add code
Jan 26, 2026
Viaarxiv icon

When Personalization Legitimizes Risks: Uncovering Safety Vulnerabilities in Personalized Dialogue Agents

Add code
Jan 25, 2026
Viaarxiv icon

OP-Bench: Benchmarking Over-Personalization for Memory-Augmented Personalized Conversational Agents

Add code
Jan 20, 2026
Viaarxiv icon

STAR-S: Improving Safety Alignment through Self-Taught Reasoning on Safety Rules

Add code
Jan 07, 2026
Viaarxiv icon

CARE-Bench: A Benchmark of Diverse Client Simulations Guided by Expert Principles for Evaluating LLMs in Psychological Counseling

Add code
Nov 12, 2025
Viaarxiv icon

Exploring and Exploiting the Inherent Efficiency within Large Reasoning Models for Self-Guided Efficiency Enhancement

Add code
Jun 18, 2025
Viaarxiv icon

Psychological Counseling Cannot Be Achieved Overnight: Automated Psychological Counseling Through Multi-Session Conversations

Add code
Jun 07, 2025
Viaarxiv icon

How Does Sequence Modeling Architecture Influence Base Capabilities of Pre-trained Language Models? Exploring Key Architecture Design Principles to Avoid Base Capabilities Degradation

Add code
May 24, 2025
Viaarxiv icon

MPO: Multilingual Safety Alignment via Reward Gap Optimization

Add code
May 22, 2025
Viaarxiv icon