Picture for Amrit Singh Bedi

Amrit Singh Bedi

Does Thinking More always Help? Understanding Test-Time Scaling in Reasoning Models

Add code
Jun 04, 2025
Viaarxiv icon

Bounded Rationality for LLMs: Satisficing Alignment at Inference-Time

Add code
May 29, 2025
Viaarxiv icon

Sample Complexity of Diffusion Model Training Without Empirical Risk Minimizer Access

Add code
May 23, 2025
Viaarxiv icon

Review, Refine, Repeat: Understanding Iterative Decoding of AI Agents with Dynamic Evaluation and Selection

Add code
Apr 02, 2025
Viaarxiv icon

Learning Multi-Robot Coordination through Locality-Based Factorized Multi-Agent Actor-Critic Algorithm

Add code
Mar 24, 2025
Viaarxiv icon

BalancedDPO: Adaptive Multi-Metric Alignment

Add code
Mar 16, 2025
Viaarxiv icon

Align-Pro: A Principled Approach to Prompt Optimization for LLM Alignment

Add code
Jan 07, 2025
Figure 1 for Align-Pro: A Principled Approach to Prompt Optimization for LLM Alignment
Figure 2 for Align-Pro: A Principled Approach to Prompt Optimization for LLM Alignment
Figure 3 for Align-Pro: A Principled Approach to Prompt Optimization for LLM Alignment
Figure 4 for Align-Pro: A Principled Approach to Prompt Optimization for LLM Alignment
Viaarxiv icon

LIAR: Leveraging Alignment (Best-of-N) to Jailbreak LLMs in Seconds

Add code
Dec 06, 2024
Figure 1 for LIAR: Leveraging Alignment (Best-of-N) to Jailbreak LLMs in Seconds
Figure 2 for LIAR: Leveraging Alignment (Best-of-N) to Jailbreak LLMs in Seconds
Figure 3 for LIAR: Leveraging Alignment (Best-of-N) to Jailbreak LLMs in Seconds
Figure 4 for LIAR: Leveraging Alignment (Best-of-N) to Jailbreak LLMs in Seconds
Viaarxiv icon

Immune: Improving Safety Against Jailbreaks in Multi-modal LLMs via Inference-Time Alignment

Add code
Nov 27, 2024
Figure 1 for Immune: Improving Safety Against Jailbreaks in Multi-modal LLMs via Inference-Time Alignment
Figure 2 for Immune: Improving Safety Against Jailbreaks in Multi-modal LLMs via Inference-Time Alignment
Figure 3 for Immune: Improving Safety Against Jailbreaks in Multi-modal LLMs via Inference-Time Alignment
Figure 4 for Immune: Improving Safety Against Jailbreaks in Multi-modal LLMs via Inference-Time Alignment
Viaarxiv icon

Hierarchical Preference Optimization: Learning to achieve goals via feasible subgoals prediction

Add code
Nov 01, 2024
Figure 1 for Hierarchical Preference Optimization: Learning to achieve goals via feasible subgoals prediction
Figure 2 for Hierarchical Preference Optimization: Learning to achieve goals via feasible subgoals prediction
Figure 3 for Hierarchical Preference Optimization: Learning to achieve goals via feasible subgoals prediction
Figure 4 for Hierarchical Preference Optimization: Learning to achieve goals via feasible subgoals prediction
Viaarxiv icon