Picture for Jizhong Han

Jizhong Han

Structured Security Auditing and Robustness Enhancement for Untrusted Agent Skills

Add code
Apr 28, 2026
Viaarxiv icon

RouteGuard: Internal-Signal Detection of Skill Poisoning in LLM Agents

Add code
Apr 24, 2026
Viaarxiv icon

FABLE: Fine-grained Fact Anchoring for Unstructured Model Editing

Add code
Apr 14, 2026
Viaarxiv icon

RecBundle: A Next-Generation Geometric Paradigm for Explainable Recommender Systems

Add code
Mar 17, 2026
Viaarxiv icon

A Cognitive Distribution and Behavior-Consistent Framework for Black-Box Attacks on Recommender Systems

Add code
Feb 12, 2026
Viaarxiv icon

Re-Align: Structured Reasoning-guided Alignment for In-Context Image Generation and Editing

Add code
Jan 08, 2026
Viaarxiv icon

Exploiting Synergistic Cognitive Biases to Bypass Safety in LLMs

Add code
Jul 30, 2025
Viaarxiv icon

Paper Summary Attack: Jailbreaking LLMs through LLM Safety Papers

Add code
Jul 17, 2025
Figure 1 for Paper Summary Attack: Jailbreaking LLMs through LLM Safety Papers
Figure 2 for Paper Summary Attack: Jailbreaking LLMs through LLM Safety Papers
Figure 3 for Paper Summary Attack: Jailbreaking LLMs through LLM Safety Papers
Figure 4 for Paper Summary Attack: Jailbreaking LLMs through LLM Safety Papers
Viaarxiv icon

Align Beyond Prompts: Evaluating World Knowledge Alignment in Text-to-Image Generation

Add code
May 24, 2025
Viaarxiv icon

LyapLock: Bounded Knowledge Preservation in Sequential Large Language Model Editing

Add code
May 21, 2025
Figure 1 for LyapLock: Bounded Knowledge Preservation in Sequential Large Language Model Editing
Figure 2 for LyapLock: Bounded Knowledge Preservation in Sequential Large Language Model Editing
Figure 3 for LyapLock: Bounded Knowledge Preservation in Sequential Large Language Model Editing
Figure 4 for LyapLock: Bounded Knowledge Preservation in Sequential Large Language Model Editing
Viaarxiv icon