Picture for Liya Su

Liya Su

Can LLMs Deeply Detect Complex Malicious Queries? A Framework for Jailbreaking via Obfuscating Intent

Add code
May 07, 2024
Viaarxiv icon

The Janus Interface: How Fine-Tuning in Large Language Models Amplifies the Privacy Risks

Add code
Oct 24, 2023
Viaarxiv icon