Alert button
Picture for Suraj Srinivas

Suraj Srinivas

Alert button

Interpreting CLIP with Sparse Linear Concept Embeddings (SpLiCE)

Add code
Bookmark button
Alert button
Feb 16, 2024
Usha Bhalla, Alex Oesterling, Suraj Srinivas, Flavio P. Calmon, Himabindu Lakkaraju

Viaarxiv icon

Certifying LLM Safety against Adversarial Prompting

Add code
Bookmark button
Alert button
Sep 06, 2023
Aounon Kumar, Chirag Agarwal, Suraj Srinivas, Soheil Feizi, Hima Lakkaraju

Viaarxiv icon

Verifiable Feature Attributions: A Bridge between Post Hoc Explainability and Inherent Interpretability

Add code
Bookmark button
Alert button
Jul 27, 2023
Usha Bhalla, Suraj Srinivas, Himabindu Lakkaraju

Figure 1 for Verifiable Feature Attributions: A Bridge between Post Hoc Explainability and Inherent Interpretability
Figure 2 for Verifiable Feature Attributions: A Bridge between Post Hoc Explainability and Inherent Interpretability
Figure 3 for Verifiable Feature Attributions: A Bridge between Post Hoc Explainability and Inherent Interpretability
Figure 4 for Verifiable Feature Attributions: A Bridge between Post Hoc Explainability and Inherent Interpretability
Viaarxiv icon

Efficient Estimation of the Local Robustness of Machine Learning Models

Add code
Bookmark button
Alert button
Jul 26, 2023
Tessa Han, Suraj Srinivas, Himabindu Lakkaraju

Figure 1 for Efficient Estimation of the Local Robustness of Machine Learning Models
Figure 2 for Efficient Estimation of the Local Robustness of Machine Learning Models
Figure 3 for Efficient Estimation of the Local Robustness of Machine Learning Models
Figure 4 for Efficient Estimation of the Local Robustness of Machine Learning Models
Viaarxiv icon

Consistent Explanations in the Face of Model Indeterminacy via Ensembling

Add code
Bookmark button
Alert button
Jun 13, 2023
Dan Ley, Leonard Tang, Matthew Nazari, Hongjin Lin, Suraj Srinivas, Himabindu Lakkaraju

Figure 1 for Consistent Explanations in the Face of Model Indeterminacy via Ensembling
Figure 2 for Consistent Explanations in the Face of Model Indeterminacy via Ensembling
Figure 3 for Consistent Explanations in the Face of Model Indeterminacy via Ensembling
Figure 4 for Consistent Explanations in the Face of Model Indeterminacy via Ensembling
Viaarxiv icon

On Minimizing the Impact of Dataset Shifts on Actionable Explanations

Add code
Bookmark button
Alert button
Jun 11, 2023
Anna P. Meyer, Dan Ley, Suraj Srinivas, Himabindu Lakkaraju

Figure 1 for On Minimizing the Impact of Dataset Shifts on Actionable Explanations
Figure 2 for On Minimizing the Impact of Dataset Shifts on Actionable Explanations
Figure 3 for On Minimizing the Impact of Dataset Shifts on Actionable Explanations
Figure 4 for On Minimizing the Impact of Dataset Shifts on Actionable Explanations
Viaarxiv icon

Word-Level Explanations for Analyzing Bias in Text-to-Image Models

Add code
Bookmark button
Alert button
Jun 03, 2023
Alexander Lin, Lucas Monteiro Paes, Sree Harsha Tanneru, Suraj Srinivas, Himabindu Lakkaraju

Figure 1 for Word-Level Explanations for Analyzing Bias in Text-to-Image Models
Figure 2 for Word-Level Explanations for Analyzing Bias in Text-to-Image Models
Figure 3 for Word-Level Explanations for Analyzing Bias in Text-to-Image Models
Figure 4 for Word-Level Explanations for Analyzing Bias in Text-to-Image Models
Viaarxiv icon

Which Models have Perceptually-Aligned Gradients? An Explanation via Off-Manifold Robustness

Add code
Bookmark button
Alert button
May 30, 2023
Suraj Srinivas, Sebastian Bordt, Hima Lakkaraju

Figure 1 for Which Models have Perceptually-Aligned Gradients? An Explanation via Off-Manifold Robustness
Figure 2 for Which Models have Perceptually-Aligned Gradients? An Explanation via Off-Manifold Robustness
Figure 3 for Which Models have Perceptually-Aligned Gradients? An Explanation via Off-Manifold Robustness
Figure 4 for Which Models have Perceptually-Aligned Gradients? An Explanation via Off-Manifold Robustness
Viaarxiv icon

Flatten the Curve: Efficiently Training Low-Curvature Neural Networks

Add code
Bookmark button
Alert button
Jun 14, 2022
Suraj Srinivas, Kyle Matoba, Himabindu Lakkaraju, Francois Fleuret

Figure 1 for Flatten the Curve: Efficiently Training Low-Curvature Neural Networks
Figure 2 for Flatten the Curve: Efficiently Training Low-Curvature Neural Networks
Figure 3 for Flatten the Curve: Efficiently Training Low-Curvature Neural Networks
Figure 4 for Flatten the Curve: Efficiently Training Low-Curvature Neural Networks
Viaarxiv icon

Which Explanation Should I Choose? A Function Approximation Perspective to Characterizing Post hoc Explanations

Add code
Bookmark button
Alert button
Jun 02, 2022
Tessa Han, Suraj Srinivas, Himabindu Lakkaraju

Figure 1 for Which Explanation Should I Choose? A Function Approximation Perspective to Characterizing Post hoc Explanations
Figure 2 for Which Explanation Should I Choose? A Function Approximation Perspective to Characterizing Post hoc Explanations
Figure 3 for Which Explanation Should I Choose? A Function Approximation Perspective to Characterizing Post hoc Explanations
Figure 4 for Which Explanation Should I Choose? A Function Approximation Perspective to Characterizing Post hoc Explanations
Viaarxiv icon