Picture for Chawin Sitawarin

Chawin Sitawarin

How much do language models memorize?

Add code
May 30, 2025
Viaarxiv icon

Lessons from Defending Gemini Against Indirect Prompt Injections

Add code
May 20, 2025
Viaarxiv icon

Vulnerability Detection with Code Language Models: How Far Are We?

Add code
Mar 27, 2024
Figure 1 for Vulnerability Detection with Code Language Models: How Far Are We?
Figure 2 for Vulnerability Detection with Code Language Models: How Far Are We?
Figure 3 for Vulnerability Detection with Code Language Models: How Far Are We?
Figure 4 for Vulnerability Detection with Code Language Models: How Far Are We?
Viaarxiv icon

PAL: Proxy-Guided Black-Box Attack on Large Language Models

Add code
Feb 15, 2024
Viaarxiv icon

Jatmo: Prompt Injection Defense by Task-Specific Finetuning

Add code
Jan 08, 2024
Viaarxiv icon

Mark My Words: Analyzing and Evaluating Language Model Watermarks

Add code
Dec 07, 2023
Figure 1 for Mark My Words: Analyzing and Evaluating Language Model Watermarks
Figure 2 for Mark My Words: Analyzing and Evaluating Language Model Watermarks
Figure 3 for Mark My Words: Analyzing and Evaluating Language Model Watermarks
Figure 4 for Mark My Words: Analyzing and Evaluating Language Model Watermarks
Viaarxiv icon

Defending Against Transfer Attacks From Public Models

Add code
Oct 26, 2023
Viaarxiv icon

OODRobustBench: benchmarking and analyzing adversarial robustness under distribution shift

Add code
Oct 19, 2023
Viaarxiv icon

SPDER: Semiperiodic Damping-Enabled Object Representation

Add code
Jun 27, 2023
Viaarxiv icon

REAP: A Large-Scale Realistic Adversarial Patch Benchmark

Add code
Dec 12, 2022
Viaarxiv icon