Adversarial Text


Adversarial text refers to a specialized text sequence that is designed specifically to influence the prediction of a language model. Generally, adversarial text attacks are carried out on large language models (LLMs). Research on understanding different adversarial approaches can help us build effective defense mechanisms to detect malicious text input and build robust language models.

FinLMM-R1: Enhancing Financial Reasoning in LMM through Scalable Data and Reward Design

Add code
Jun 16, 2025
Viaarxiv icon

Navigating the Black Box: Leveraging LLMs for Effective Text-Level Graph Injection Attacks

Add code
Jun 16, 2025
Viaarxiv icon

NAP-Tuning: Neural Augmented Prompt Tuning for Adversarially Robust Vision-Language Models

Add code
Jun 15, 2025
Viaarxiv icon

Transforming Chatbot Text: A Sequence-to-Sequence Approach

Add code
Jun 15, 2025
Viaarxiv icon

Pushing the Limits of Safety: A Technical Report on the ATLAS Challenge 2025

Add code
Jun 14, 2025
Viaarxiv icon

TrustGLM: Evaluating the Robustness of GraphLLMs Against Prompt, Text, and Structure Attacks

Add code
Jun 13, 2025
Viaarxiv icon

Unsourced Adversarial CAPTCHA: A Bi-Phase Adversarial CAPTCHA Framework

Add code
Jun 12, 2025
Viaarxiv icon

Adversarial Text Generation with Dynamic Contextual Perturbation

Add code
Jun 10, 2025
Viaarxiv icon

AngleRoCL: Angle-Robust Concept Learning for Physically View-Invariant T2I Adversarial Patches

Add code
Jun 11, 2025
Viaarxiv icon

GenBreak: Red Teaming Text-to-Image Generators Using Large Language Models

Add code
Jun 11, 2025
Viaarxiv icon