Picture for Ilia Shumailov

Ilia Shumailov

Cascading Adversarial Bias from Injection to Distillation in Language Models

Add code
May 30, 2025
Viaarxiv icon

Machine Learning Models Have a Supply Chain Problem

Add code
May 28, 2025
Viaarxiv icon

Strong Membership Inference Attacks on Massive Datasets and (Moderately) Large Language Models

Add code
May 24, 2025
Viaarxiv icon

Architectural Backdoors for Within-Batch Data Stealing and Model Inference Manipulation

Add code
May 23, 2025
Viaarxiv icon

Lessons from Defending Gemini Against Indirect Prompt Injections

Add code
May 20, 2025
Viaarxiv icon

Watermarking Needs Input Repetition Masking

Add code
Apr 16, 2025
Viaarxiv icon

Defeating Prompt Injections by Design

Add code
Mar 24, 2025
Viaarxiv icon

Interpreting the Repeated Token Phenomenon in Large Language Models

Add code
Mar 11, 2025
Viaarxiv icon

Trusted Machine Learning Models Unlock Private Inference for Problems Currently Infeasible with Cryptography

Add code
Jan 15, 2025
Viaarxiv icon

Machine Unlearning Doesn't Do What You Think: Lessons for Generative AI Policy, Research, and Practice

Add code
Dec 09, 2024
Figure 1 for Machine Unlearning Doesn't Do What You Think: Lessons for Generative AI Policy, Research, and Practice
Figure 2 for Machine Unlearning Doesn't Do What You Think: Lessons for Generative AI Policy, Research, and Practice
Figure 3 for Machine Unlearning Doesn't Do What You Think: Lessons for Generative AI Policy, Research, and Practice
Figure 4 for Machine Unlearning Doesn't Do What You Think: Lessons for Generative AI Policy, Research, and Practice
Viaarxiv icon