Adversarial


Engineering Robustness into Personal Agents with the AI Workflow Store

Add code
May 11, 2026
Viaarxiv icon

RUBEN: Rule-Based Explanations for Retrieval-Augmented LLM Systems

Add code
May 11, 2026
Viaarxiv icon

Benchmarking Sensor-Fault Robustness in Forecasting

Add code
May 11, 2026
Viaarxiv icon

Mistake-Bounded Language Generation

Add code
May 11, 2026
Viaarxiv icon

LLMs for Secure Hardware Design and Related Problems: Opportunities and Challenges

Add code
May 11, 2026
Viaarxiv icon

Fixed-Point Neural Optimal Transport without Implicit Differentiation

Add code
May 11, 2026
Viaarxiv icon

LITMUS: Benchmarking Behavioral Jailbreaks of LLM Agents in Real OS Environments

Add code
May 11, 2026
Viaarxiv icon

Conformity Generates Collective Misalignment in AI Agents Societies

Add code
May 11, 2026
Viaarxiv icon

Not Blind but Silenced: Rebalancing Vision and Language via Adversarial Counter-Commonsense Equilibrium

Add code
May 11, 2026
Viaarxiv icon

PRISM: Generation-Time Detection and Mitigation of Secret Leakage in Multi-Agent LLM Pipelines

Add code
May 11, 2026
Viaarxiv icon