Picture for Yossi Gandelsman

Yossi Gandelsman

Same Task, Different Circuits: Disentangling Modality-Specific Mechanisms in VLMs

Add code
Jun 11, 2025
Viaarxiv icon

Vision Transformers Don't Need Trained Registers

Add code
Jun 09, 2025
Figure 1 for Vision Transformers Don't Need Trained Registers
Figure 2 for Vision Transformers Don't Need Trained Registers
Figure 3 for Vision Transformers Don't Need Trained Registers
Figure 4 for Vision Transformers Don't Need Trained Registers
Viaarxiv icon

Steering CLIP's vision transformer with sparse autoencoders

Add code
Apr 11, 2025
Figure 1 for Steering CLIP's vision transformer with sparse autoencoders
Figure 2 for Steering CLIP's vision transformer with sparse autoencoders
Figure 3 for Steering CLIP's vision transformer with sparse autoencoders
Figure 4 for Steering CLIP's vision transformer with sparse autoencoders
Viaarxiv icon

Teaching Humans Subtle Differences with DIFFusion

Add code
Apr 10, 2025
Figure 1 for Teaching Humans Subtle Differences with DIFFusion
Figure 2 for Teaching Humans Subtle Differences with DIFFusion
Figure 3 for Teaching Humans Subtle Differences with DIFFusion
Figure 4 for Teaching Humans Subtle Differences with DIFFusion
Viaarxiv icon

Interpreting the Repeated Token Phenomenon in Large Language Models

Add code
Mar 11, 2025
Figure 1 for Interpreting the Repeated Token Phenomenon in Large Language Models
Figure 2 for Interpreting the Repeated Token Phenomenon in Large Language Models
Figure 3 for Interpreting the Repeated Token Phenomenon in Large Language Models
Figure 4 for Interpreting the Repeated Token Phenomenon in Large Language Models
Viaarxiv icon

LLMs can see and hear without any training

Add code
Jan 30, 2025
Figure 1 for LLMs can see and hear without any training
Figure 2 for LLMs can see and hear without any training
Figure 3 for LLMs can see and hear without any training
Figure 4 for LLMs can see and hear without any training
Viaarxiv icon

An Empirical Study of Autoregressive Pre-training from Videos

Add code
Jan 09, 2025
Figure 1 for An Empirical Study of Autoregressive Pre-training from Videos
Figure 2 for An Empirical Study of Autoregressive Pre-training from Videos
Figure 3 for An Empirical Study of Autoregressive Pre-training from Videos
Figure 4 for An Empirical Study of Autoregressive Pre-training from Videos
Viaarxiv icon

Learning Video Representations without Natural Videos

Add code
Oct 31, 2024
Figure 1 for Learning Video Representations without Natural Videos
Figure 2 for Learning Video Representations without Natural Videos
Figure 3 for Learning Video Representations without Natural Videos
Figure 4 for Learning Video Representations without Natural Videos
Viaarxiv icon

Interpreting and Editing Vision-Language Representations to Mitigate Hallucinations

Add code
Oct 03, 2024
Figure 1 for Interpreting and Editing Vision-Language Representations to Mitigate Hallucinations
Figure 2 for Interpreting and Editing Vision-Language Representations to Mitigate Hallucinations
Figure 3 for Interpreting and Editing Vision-Language Representations to Mitigate Hallucinations
Figure 4 for Interpreting and Editing Vision-Language Representations to Mitigate Hallucinations
Viaarxiv icon

Quantifying and Enabling the Interpretability of CLIP-like Models

Add code
Sep 10, 2024
Viaarxiv icon