Data Poisoning


Data poisoning is the process of manipulating training data to compromise the performance of machine learning models.

The Trigger in the Haystack: Extracting and Reconstructing LLM Backdoor Triggers

Add code
Feb 03, 2026
Viaarxiv icon

When Attention Betrays: Erasing Backdoor Attacks in Robotic Policies by Reconstructing Visual Tokens

Add code
Feb 03, 2026
Viaarxiv icon

Human Society-Inspired Approaches to Agentic AI Security: The 4C Framework

Add code
Feb 02, 2026
Viaarxiv icon

Trustworthy Blockchain-based Federated Learning for Electronic Health Records: Securing Participant Identity with Decentralized Identifiers and Verifiable Credentials

Add code
Feb 02, 2026
Viaarxiv icon

Safety-Efficacy Trade Off: Robustness against Data-Poisoning

Add code
Jan 31, 2026
Viaarxiv icon

TinyGuard:A lightweight Byzantine Defense for Resource-Constrained Federated Learning via Statistical Update Fingerprints

Add code
Feb 02, 2026
Viaarxiv icon

TCAP: Tri-Component Attention Profiling for Unsupervised Backdoor Detection in MLLM Fine-Tuning

Add code
Jan 29, 2026
Viaarxiv icon

RPP: A Certified Poisoned-Sample Detection Framework for Backdoor Attacks under Dataset Imbalance

Add code
Jan 30, 2026
Viaarxiv icon

Thought-Transfer: Indirect Targeted Poisoning Attacks on Chain-of-Thought Reasoning Models

Add code
Jan 27, 2026
Viaarxiv icon

Do LLMs Truly Benefit from Longer Context in Automatic Post-Editing?

Add code
Jan 27, 2026
Viaarxiv icon