Picture for Thomas Fel

Thomas Fel

IRIT

Evaluating Sparse Autoencoders: From Shallow Design to Matching Pursuit

Add code
Jun 05, 2025
Viaarxiv icon

Interpreting the Linear Structure of Vision-language Model Embedding Spaces

Add code
Apr 16, 2025
Viaarxiv icon

Projecting Assumptions: The Duality Between Sparse Autoencoders and Concept Geometry

Add code
Mar 03, 2025
Viaarxiv icon

Archetypal SAE: Adaptive and Stable Dictionary Learning for Concept Extraction in Large Vision Models

Add code
Feb 18, 2025
Viaarxiv icon

Universal Sparse Autoencoders: Interpretable Cross-Model Concept Alignment

Add code
Feb 06, 2025
Viaarxiv icon

Sparks of Explainability: Recent Advancements in Explaining Large Vision Models

Add code
Feb 03, 2025
Viaarxiv icon

An Adaptive Orthogonal Convolution Scheme for Efficient and Flexible CNN Architectures

Add code
Jan 14, 2025
Viaarxiv icon

Local vs distributed representations: What is the right basis for interpretability?

Add code
Nov 06, 2024
Viaarxiv icon

Unearthing Skill-Level Insights for Understanding Trade-Offs of Foundation Models

Add code
Oct 17, 2024
Figure 1 for Unearthing Skill-Level Insights for Understanding Trade-Offs of Foundation Models
Figure 2 for Unearthing Skill-Level Insights for Understanding Trade-Offs of Foundation Models
Figure 3 for Unearthing Skill-Level Insights for Understanding Trade-Offs of Foundation Models
Figure 4 for Unearthing Skill-Level Insights for Understanding Trade-Offs of Foundation Models
Viaarxiv icon

One Wave to Explain Them All: A Unifying Perspective on Post-hoc Explainability

Add code
Oct 02, 2024
Viaarxiv icon