Picture for Jiaheng Zhang

Jiaheng Zhang

Efficient Reasoning via Chain of Unconscious Thought

Add code
May 26, 2025
Viaarxiv icon

Silent Leaks: Implicit Knowledge Extraction Attack on RAG Systems through Benign Queries

Add code
May 21, 2025
Viaarxiv icon

GuardReasoner-VL: Safeguarding VLMs via Reinforced Reasoning

Add code
May 16, 2025
Viaarxiv icon

Geneshift: Impact of different scenario shift on Jailbreaking LLM

Add code
Apr 10, 2025
Viaarxiv icon

Efficient Inference for Large Reasoning Models: A Survey

Add code
Mar 29, 2025
Viaarxiv icon

Sparse Autoencoder as a Zero-Shot Classifier for Concept Erasing in Text-to-Image Diffusion Models

Add code
Mar 12, 2025
Viaarxiv icon

Navigating the Helpfulness-Truthfulness Trade-Off with Uncertainty-Aware Instruction Fine-Tuning

Add code
Feb 17, 2025
Viaarxiv icon

GuardReasoner: Towards Reasoning-based LLM Safeguards

Add code
Jan 30, 2025
Figure 1 for GuardReasoner: Towards Reasoning-based LLM Safeguards
Figure 2 for GuardReasoner: Towards Reasoning-based LLM Safeguards
Figure 3 for GuardReasoner: Towards Reasoning-based LLM Safeguards
Figure 4 for GuardReasoner: Towards Reasoning-based LLM Safeguards
Viaarxiv icon

Political-LLM: Large Language Models in Political Science

Add code
Dec 09, 2024
Figure 1 for Political-LLM: Large Language Models in Political Science
Figure 2 for Political-LLM: Large Language Models in Political Science
Figure 3 for Political-LLM: Large Language Models in Political Science
Figure 4 for Political-LLM: Large Language Models in Political Science
Viaarxiv icon

Lazarus: Resilient and Elastic Training of Mixture-of-Experts Models with Adaptive Expert Placement

Add code
Jul 05, 2024
Figure 1 for Lazarus: Resilient and Elastic Training of Mixture-of-Experts Models with Adaptive Expert Placement
Figure 2 for Lazarus: Resilient and Elastic Training of Mixture-of-Experts Models with Adaptive Expert Placement
Figure 3 for Lazarus: Resilient and Elastic Training of Mixture-of-Experts Models with Adaptive Expert Placement
Figure 4 for Lazarus: Resilient and Elastic Training of Mixture-of-Experts Models with Adaptive Expert Placement
Viaarxiv icon