Picture for Supriyo Chakraborty

Supriyo Chakraborty

IBM Research

Your Model Diversity, Not Method, Determines Reasoning Strategy

Add code
Apr 12, 2026
Viaarxiv icon

Decomposing the Delta: What Do Models Actually Learn from Preference Pairs?

Add code
Apr 09, 2026
Viaarxiv icon

Routing with Generated Data: Annotation-Free LLM Skill Estimation and Expert Selection

Add code
Jan 14, 2026
Viaarxiv icon

Leveraging Parameter Space Symmetries for Reasoning Skill Transfer in LLMs

Add code
Nov 13, 2025
Viaarxiv icon

SPEAR-MM: Selective Parameter Evaluation and Restoration via Model Merging for Efficient Financial LLM Adaptation

Add code
Nov 11, 2025
Figure 1 for SPEAR-MM: Selective Parameter Evaluation and Restoration via Model Merging for Efficient Financial LLM Adaptation
Figure 2 for SPEAR-MM: Selective Parameter Evaluation and Restoration via Model Merging for Efficient Financial LLM Adaptation
Figure 3 for SPEAR-MM: Selective Parameter Evaluation and Restoration via Model Merging for Efficient Financial LLM Adaptation
Figure 4 for SPEAR-MM: Selective Parameter Evaluation and Restoration via Model Merging for Efficient Financial LLM Adaptation
Viaarxiv icon

Optimizing Reasoning Efficiency through Prompt Difficulty Prediction

Add code
Nov 05, 2025
Viaarxiv icon

On Understanding of the Dynamics of Model Capacity in Continual Learning

Add code
Aug 11, 2025
Viaarxiv icon

Training Dynamics Underlying Language Model Scaling Laws: Loss Deceleration and Zero-Sum Learning

Add code
Jun 05, 2025
Viaarxiv icon

Critique-Guided Distillation: Improving Supervised Fine-tuning via Better Distillation

Add code
May 16, 2025
Figure 1 for Critique-Guided Distillation: Improving Supervised Fine-tuning via Better Distillation
Figure 2 for Critique-Guided Distillation: Improving Supervised Fine-tuning via Better Distillation
Figure 3 for Critique-Guided Distillation: Improving Supervised Fine-tuning via Better Distillation
Figure 4 for Critique-Guided Distillation: Improving Supervised Fine-tuning via Better Distillation
Viaarxiv icon

Dense Backpropagation Improves Training for Sparse Mixture-of-Experts

Add code
Apr 18, 2025
Viaarxiv icon