Picture for Giulio Zizzo

Giulio Zizzo

TRACES: Tagging Reasoning Steps for Adaptive Cost-Efficient Early-Stopping

Add code
Apr 22, 2026
Viaarxiv icon

Breaking MCP with Function Hijacking Attacks: Novel Threats for Function Calling and Agentic Models

Add code
Apr 22, 2026
Viaarxiv icon

Blue Teaming Function-Calling Agents

Add code
Jan 14, 2026
Viaarxiv icon

Step-Tagging: Toward controlling the generation of Language Reasoning Models through step monitoring

Add code
Dec 16, 2025
Viaarxiv icon

Pre-Hoc Predictions in AutoML: Leveraging LLMs to Enhance Model Selection and Benchmarking for Tabular datasets

Add code
Oct 02, 2025
Viaarxiv icon

Activated LoRA: Fine-tuned LLMs for Intrinsics

Add code
Apr 16, 2025
Viaarxiv icon

MAD-MAX: Modular And Diverse Malicious Attack MiXtures for Automated LLM Red Teaming

Add code
Mar 08, 2025
Figure 1 for MAD-MAX: Modular And Diverse Malicious Attack MiXtures for Automated LLM Red Teaming
Figure 2 for MAD-MAX: Modular And Diverse Malicious Attack MiXtures for Automated LLM Red Teaming
Figure 3 for MAD-MAX: Modular And Diverse Malicious Attack MiXtures for Automated LLM Red Teaming
Figure 4 for MAD-MAX: Modular And Diverse Malicious Attack MiXtures for Automated LLM Red Teaming
Viaarxiv icon

Adversarial Prompt Evaluation: Systematic Benchmarking of Guardrails Against Prompt Input Attacks on LLMs

Add code
Feb 21, 2025
Figure 1 for Adversarial Prompt Evaluation: Systematic Benchmarking of Guardrails Against Prompt Input Attacks on LLMs
Figure 2 for Adversarial Prompt Evaluation: Systematic Benchmarking of Guardrails Against Prompt Input Attacks on LLMs
Figure 3 for Adversarial Prompt Evaluation: Systematic Benchmarking of Guardrails Against Prompt Input Attacks on LLMs
Figure 4 for Adversarial Prompt Evaluation: Systematic Benchmarking of Guardrails Against Prompt Input Attacks on LLMs
Viaarxiv icon

Granite Guardian

Add code
Dec 10, 2024
Figure 1 for Granite Guardian
Figure 2 for Granite Guardian
Figure 3 for Granite Guardian
Figure 4 for Granite Guardian
Viaarxiv icon

HarmLevelBench: Evaluating Harm-Level Compliance and the Impact of Quantization on Model Alignment

Add code
Nov 11, 2024
Viaarxiv icon