Picture for Ramtin Pedarsani

Ramtin Pedarsani

Deactivating Refusal Triggers: Understanding and Mitigating Overrefusal in Safety Alignment

Add code
Mar 12, 2026
Viaarxiv icon

MMLoP: Multi-Modal Low-Rank Prompting for Efficient Vision-Language Adaptation

Add code
Feb 24, 2026
Viaarxiv icon

Enhancing the Safety of Medical Vision-Language Models by Synthetic Demonstrations

Add code
Jun 08, 2025
Viaarxiv icon

Few-Shot Adversarial Low-Rank Fine-Tuning of Vision-Language Models

Add code
May 21, 2025
Viaarxiv icon

The Safety-Privacy Tradeoff in Linear Bandits

Add code
Apr 23, 2025
Viaarxiv icon

SPEX: Scaling Feature Interaction Explanations for LLMs

Add code
Feb 19, 2025
Figure 1 for SPEX: Scaling Feature Interaction Explanations for LLMs
Figure 2 for SPEX: Scaling Feature Interaction Explanations for LLMs
Figure 3 for SPEX: Scaling Feature Interaction Explanations for LLMs
Figure 4 for SPEX: Scaling Feature Interaction Explanations for LLMs
Viaarxiv icon

Decentralized Low-Rank Fine-Tuning of Large Language Models

Add code
Jan 26, 2025
Figure 1 for Decentralized Low-Rank Fine-Tuning of Large Language Models
Figure 2 for Decentralized Low-Rank Fine-Tuning of Large Language Models
Figure 3 for Decentralized Low-Rank Fine-Tuning of Large Language Models
Figure 4 for Decentralized Low-Rank Fine-Tuning of Large Language Models
Viaarxiv icon

No Free Lunch for Defending Against Prefilling Attack by In-Context Learning

Add code
Dec 13, 2024
Viaarxiv icon

Multi-Bin Batching for Increasing LLM Inference Throughput

Add code
Dec 03, 2024
Figure 1 for Multi-Bin Batching for Increasing LLM Inference Throughput
Figure 2 for Multi-Bin Batching for Increasing LLM Inference Throughput
Figure 3 for Multi-Bin Batching for Increasing LLM Inference Throughput
Figure 4 for Multi-Bin Batching for Increasing LLM Inference Throughput
Viaarxiv icon

Conflict-Aware Adversarial Training

Add code
Oct 21, 2024
Viaarxiv icon