Picture for Wojciech Samek

Wojciech Samek

The Atlas of In-Context Learning: How Attention Heads Shape In-Context Retrieval Augmentation

Add code
May 21, 2025
Viaarxiv icon

Steering CLIP's vision transformer with sparse autoencoders

Add code
Apr 11, 2025
Figure 1 for Steering CLIP's vision transformer with sparse autoencoders
Figure 2 for Steering CLIP's vision transformer with sparse autoencoders
Figure 3 for Steering CLIP's vision transformer with sparse autoencoders
Figure 4 for Steering CLIP's vision transformer with sparse autoencoders
Viaarxiv icon

Efficient Federated Learning Tiny Language Models for Mobile Network Feature Prediction

Add code
Apr 02, 2025
Viaarxiv icon

ASIDE: Architectural Separation of Instructions and Data in Language Models

Add code
Mar 13, 2025
Figure 1 for ASIDE: Architectural Separation of Instructions and Data in Language Models
Figure 2 for ASIDE: Architectural Separation of Instructions and Data in Language Models
Figure 3 for ASIDE: Architectural Separation of Instructions and Data in Language Models
Figure 4 for ASIDE: Architectural Separation of Instructions and Data in Language Models
Viaarxiv icon

Post-Hoc Concept Disentanglement: From Correlated to Isolated Concept Representations

Add code
Mar 07, 2025
Viaarxiv icon

FADE: Why Bad Descriptions Happen to Good Features

Add code
Feb 24, 2025
Viaarxiv icon

A Close Look at Decomposition-based XAI-Methods for Transformer Language Models

Add code
Feb 21, 2025
Viaarxiv icon

Synthetic Datasets for Machine Learning on Spatio-Temporal Graphs using PDEs

Add code
Feb 06, 2025
Viaarxiv icon

Ensuring Medical AI Safety: Explainable AI-Driven Detection and Mitigation of Spurious Model Behavior and Associated Data

Add code
Jan 23, 2025
Figure 1 for Ensuring Medical AI Safety: Explainable AI-Driven Detection and Mitigation of Spurious Model Behavior and Associated Data
Figure 2 for Ensuring Medical AI Safety: Explainable AI-Driven Detection and Mitigation of Spurious Model Behavior and Associated Data
Figure 3 for Ensuring Medical AI Safety: Explainable AI-Driven Detection and Mitigation of Spurious Model Behavior and Associated Data
Figure 4 for Ensuring Medical AI Safety: Explainable AI-Driven Detection and Mitigation of Spurious Model Behavior and Associated Data
Viaarxiv icon

Mechanistic understanding and validation of large AI models with SemanticLens

Add code
Jan 09, 2025
Figure 1 for Mechanistic understanding and validation of large AI models with SemanticLens
Figure 2 for Mechanistic understanding and validation of large AI models with SemanticLens
Viaarxiv icon