Picture for Saurav Jha

Saurav Jha

Reconstruction or Semantics? What Makes a Latent Space Useful for Robotic World Models

Add code
May 07, 2026
Viaarxiv icon

REAM: Merging Improves Pruning of Experts in LLMs

Add code
Apr 06, 2026
Viaarxiv icon

Lightweight Structured Multimodal Reasoning for Clinical Scene Understanding in Robotics

Add code
Sep 26, 2025
Figure 1 for Lightweight Structured Multimodal Reasoning for Clinical Scene Understanding in Robotics
Figure 2 for Lightweight Structured Multimodal Reasoning for Clinical Scene Understanding in Robotics
Figure 3 for Lightweight Structured Multimodal Reasoning for Clinical Scene Understanding in Robotics
Figure 4 for Lightweight Structured Multimodal Reasoning for Clinical Scene Understanding in Robotics
Viaarxiv icon

GLOV: Guided Large Language Models as Implicit Optimizers for Vision Language Models

Add code
Oct 08, 2024
Figure 1 for GLOV: Guided Large Language Models as Implicit Optimizers for Vision Language Models
Figure 2 for GLOV: Guided Large Language Models as Implicit Optimizers for Vision Language Models
Figure 3 for GLOV: Guided Large Language Models as Implicit Optimizers for Vision Language Models
Figure 4 for GLOV: Guided Large Language Models as Implicit Optimizers for Vision Language Models
Viaarxiv icon

Mining Your Own Secrets: Diffusion Classifier Scores for Continual Personalization of Text-to-Image Diffusion Models

Add code
Oct 02, 2024
Figure 1 for Mining Your Own Secrets: Diffusion Classifier Scores for Continual Personalization of Text-to-Image Diffusion Models
Figure 2 for Mining Your Own Secrets: Diffusion Classifier Scores for Continual Personalization of Text-to-Image Diffusion Models
Figure 3 for Mining Your Own Secrets: Diffusion Classifier Scores for Continual Personalization of Text-to-Image Diffusion Models
Figure 4 for Mining Your Own Secrets: Diffusion Classifier Scores for Continual Personalization of Text-to-Image Diffusion Models
Viaarxiv icon

On the relevance of pre-neural approaches in natural language processing pedagogy

Add code
May 16, 2024
Figure 1 for On the relevance of pre-neural approaches in natural language processing pedagogy
Figure 2 for On the relevance of pre-neural approaches in natural language processing pedagogy
Figure 3 for On the relevance of pre-neural approaches in natural language processing pedagogy
Figure 4 for On the relevance of pre-neural approaches in natural language processing pedagogy
Viaarxiv icon

CLAP4CLIP: Continual Learning with Probabilistic Finetuning for Vision-Language Models

Add code
Mar 28, 2024
Figure 1 for CLAP4CLIP: Continual Learning with Probabilistic Finetuning for Vision-Language Models
Figure 2 for CLAP4CLIP: Continual Learning with Probabilistic Finetuning for Vision-Language Models
Figure 3 for CLAP4CLIP: Continual Learning with Probabilistic Finetuning for Vision-Language Models
Figure 4 for CLAP4CLIP: Continual Learning with Probabilistic Finetuning for Vision-Language Models
Viaarxiv icon

NPCL: Neural Processes for Uncertainty-Aware Continual Learning

Add code
Oct 30, 2023
Figure 1 for NPCL: Neural Processes for Uncertainty-Aware Continual Learning
Figure 2 for NPCL: Neural Processes for Uncertainty-Aware Continual Learning
Figure 3 for NPCL: Neural Processes for Uncertainty-Aware Continual Learning
Figure 4 for NPCL: Neural Processes for Uncertainty-Aware Continual Learning
Viaarxiv icon

Distilled Reverse Attention Network for Open-world Compositional Zero-Shot Learning

Add code
Mar 01, 2023
Figure 1 for Distilled Reverse Attention Network for Open-world Compositional Zero-Shot Learning
Figure 2 for Distilled Reverse Attention Network for Open-world Compositional Zero-Shot Learning
Figure 3 for Distilled Reverse Attention Network for Open-world Compositional Zero-Shot Learning
Figure 4 for Distilled Reverse Attention Network for Open-world Compositional Zero-Shot Learning
Viaarxiv icon

Towards Exemplar-Free Continual Learning in Vision Transformers: an Account of Attention, Functional and Weight Regularization

Add code
Mar 28, 2022
Figure 1 for Towards Exemplar-Free Continual Learning in Vision Transformers: an Account of Attention, Functional and Weight Regularization
Figure 2 for Towards Exemplar-Free Continual Learning in Vision Transformers: an Account of Attention, Functional and Weight Regularization
Figure 3 for Towards Exemplar-Free Continual Learning in Vision Transformers: an Account of Attention, Functional and Weight Regularization
Figure 4 for Towards Exemplar-Free Continual Learning in Vision Transformers: an Account of Attention, Functional and Weight Regularization
Viaarxiv icon