Picture for Udari Madhushani Sehwag

Udari Madhushani Sehwag

ROK-FORTRESS: Measuring the Effect of Geopolitical Transcreation for National Security and Public Safety

Add code
May 13, 2026
Viaarxiv icon

SciPredict: Can LLMs Predict the Outcomes of Scientific Experiments in Natural Sciences?

Add code
Apr 12, 2026
Viaarxiv icon

Defensive Refusal Bias: How Safety Alignment Fails Cyber Defenders

Add code
Mar 01, 2026
Viaarxiv icon

LHAW: Controllable Underspecification for Long-Horizon Tasks

Add code
Feb 11, 2026
Viaarxiv icon

Best Practices for Biorisk Evaluations on Open-Weight Bio-Foundation Models

Add code
Nov 03, 2025
Viaarxiv icon

Remote Labor Index: Measuring AI Automation of Remote Work

Add code
Oct 30, 2025
Viaarxiv icon

Collab: Controlled Decoding using Mixture of Agents for LLM Alignment

Add code
Mar 27, 2025
Viaarxiv icon

In-Context Learning with Topological Information for Knowledge Graph Completion

Add code
Dec 11, 2024
Viaarxiv icon

AdvBDGen: Adversarially Fortified Prompt-Specific Fuzzy Backdoor Generator Against LLM Alignment

Add code
Oct 15, 2024
Figure 1 for AdvBDGen: Adversarially Fortified Prompt-Specific Fuzzy Backdoor Generator Against LLM Alignment
Figure 2 for AdvBDGen: Adversarially Fortified Prompt-Specific Fuzzy Backdoor Generator Against LLM Alignment
Figure 3 for AdvBDGen: Adversarially Fortified Prompt-Specific Fuzzy Backdoor Generator Against LLM Alignment
Figure 4 for AdvBDGen: Adversarially Fortified Prompt-Specific Fuzzy Backdoor Generator Against LLM Alignment
Viaarxiv icon

Can LLMs be Scammed? A Baseline Measurement Study

Add code
Oct 14, 2024
Figure 1 for Can LLMs be Scammed? A Baseline Measurement Study
Figure 2 for Can LLMs be Scammed? A Baseline Measurement Study
Figure 3 for Can LLMs be Scammed? A Baseline Measurement Study
Figure 4 for Can LLMs be Scammed? A Baseline Measurement Study
Viaarxiv icon