Picture for Lihai Nie

Lihai Nie

Benchmarking Poisoning Attacks against Retrieval-Augmented Generation

Add code
May 24, 2025
Viaarxiv icon

When Safety Detectors Aren't Enough: A Stealthy and Effective Jailbreak Attack on LLMs via Steganographic Techniques

Add code
May 22, 2025
Viaarxiv icon

CTRAP: Embedding Collapse Trap to Safeguard Large Language Models from Harmful Fine-Tuning

Add code
May 22, 2025
Viaarxiv icon

Practical Poisoning Attacks against Retrieval-Augmented Generation

Add code
Apr 04, 2025
Viaarxiv icon

Your Fixed Watermark is Fragile: Towards Semantic-Aware Watermark for EaaS Copyright Protection

Add code
Nov 14, 2024
Viaarxiv icon