Picture for Yossi Gandelsman

Yossi Gandelsman

Steering CLIP's vision transformer with sparse autoencoders

Add code
Apr 11, 2025
Viaarxiv icon

Teaching Humans Subtle Differences with DIFFusion

Add code
Apr 10, 2025
Viaarxiv icon

Interpreting the Repeated Token Phenomenon in Large Language Models

Add code
Mar 11, 2025
Viaarxiv icon

LLMs can see and hear without any training

Add code
Jan 30, 2025
Viaarxiv icon

An Empirical Study of Autoregressive Pre-training from Videos

Add code
Jan 09, 2025
Figure 1 for An Empirical Study of Autoregressive Pre-training from Videos
Figure 2 for An Empirical Study of Autoregressive Pre-training from Videos
Figure 3 for An Empirical Study of Autoregressive Pre-training from Videos
Figure 4 for An Empirical Study of Autoregressive Pre-training from Videos
Viaarxiv icon

Learning Video Representations without Natural Videos

Add code
Oct 31, 2024
Viaarxiv icon

Interpreting and Editing Vision-Language Representations to Mitigate Hallucinations

Add code
Oct 03, 2024
Figure 1 for Interpreting and Editing Vision-Language Representations to Mitigate Hallucinations
Figure 2 for Interpreting and Editing Vision-Language Representations to Mitigate Hallucinations
Figure 3 for Interpreting and Editing Vision-Language Representations to Mitigate Hallucinations
Figure 4 for Interpreting and Editing Vision-Language Representations to Mitigate Hallucinations
Viaarxiv icon

Quantifying and Enabling the Interpretability of CLIP-like Models

Add code
Sep 10, 2024
Viaarxiv icon

Interpreting the Weight Space of Customized Diffusion Models

Add code
Jun 13, 2024
Viaarxiv icon

Interpreting the Second-Order Effects of Neurons in CLIP

Add code
Jun 06, 2024
Viaarxiv icon