Alert button
Picture for Himabindu Lakkaraju

Himabindu Lakkaraju

Alert button

In-Context Unlearning: Language Models as Few Shot Unlearners

Add code
Bookmark button
Alert button
Oct 12, 2023
Martin Pawelczyk, Seth Neel, Himabindu Lakkaraju

Figure 1 for In-Context Unlearning: Language Models as Few Shot Unlearners
Figure 2 for In-Context Unlearning: Language Models as Few Shot Unlearners
Figure 3 for In-Context Unlearning: Language Models as Few Shot Unlearners
Figure 4 for In-Context Unlearning: Language Models as Few Shot Unlearners
Viaarxiv icon

Are Large Language Models Post Hoc Explainers?

Add code
Bookmark button
Alert button
Oct 10, 2023
Nicholas Kroeger, Dan Ley, Satyapriya Krishna, Chirag Agarwal, Himabindu Lakkaraju

Figure 1 for Are Large Language Models Post Hoc Explainers?
Figure 2 for Are Large Language Models Post Hoc Explainers?
Figure 3 for Are Large Language Models Post Hoc Explainers?
Figure 4 for Are Large Language Models Post Hoc Explainers?
Viaarxiv icon

On the Trade-offs between Adversarial Robustness and Actionable Explanations

Add code
Bookmark button
Alert button
Sep 28, 2023
Satyapriya Krishna, Chirag Agarwal, Himabindu Lakkaraju

Viaarxiv icon

Accurate, Explainable, and Private Models: Providing Recourse While Minimizing Training Data Leakage

Add code
Bookmark button
Alert button
Aug 08, 2023
Catherine Huang, Chelse Swoopes, Christina Xiao, Jiaqi Ma, Himabindu Lakkaraju

Figure 1 for Accurate, Explainable, and Private Models: Providing Recourse While Minimizing Training Data Leakage
Figure 2 for Accurate, Explainable, and Private Models: Providing Recourse While Minimizing Training Data Leakage
Figure 3 for Accurate, Explainable, and Private Models: Providing Recourse While Minimizing Training Data Leakage
Figure 4 for Accurate, Explainable, and Private Models: Providing Recourse While Minimizing Training Data Leakage
Viaarxiv icon

Verifiable Feature Attributions: A Bridge between Post Hoc Explainability and Inherent Interpretability

Add code
Bookmark button
Alert button
Jul 27, 2023
Usha Bhalla, Suraj Srinivas, Himabindu Lakkaraju

Figure 1 for Verifiable Feature Attributions: A Bridge between Post Hoc Explainability and Inherent Interpretability
Figure 2 for Verifiable Feature Attributions: A Bridge between Post Hoc Explainability and Inherent Interpretability
Figure 3 for Verifiable Feature Attributions: A Bridge between Post Hoc Explainability and Inherent Interpretability
Figure 4 for Verifiable Feature Attributions: A Bridge between Post Hoc Explainability and Inherent Interpretability
Viaarxiv icon

Efficient Estimation of the Local Robustness of Machine Learning Models

Add code
Bookmark button
Alert button
Jul 26, 2023
Tessa Han, Suraj Srinivas, Himabindu Lakkaraju

Figure 1 for Efficient Estimation of the Local Robustness of Machine Learning Models
Figure 2 for Efficient Estimation of the Local Robustness of Machine Learning Models
Figure 3 for Efficient Estimation of the Local Robustness of Machine Learning Models
Figure 4 for Efficient Estimation of the Local Robustness of Machine Learning Models
Viaarxiv icon

Analyzing Chain-of-Thought Prompting in Large Language Models via Gradient-based Feature Attributions

Add code
Bookmark button
Alert button
Jul 25, 2023
Skyler Wu, Eric Meng Shen, Charumathi Badrinath, Jiaqi Ma, Himabindu Lakkaraju

Figure 1 for Analyzing Chain-of-Thought Prompting in Large Language Models via Gradient-based Feature Attributions
Figure 2 for Analyzing Chain-of-Thought Prompting in Large Language Models via Gradient-based Feature Attributions
Figure 3 for Analyzing Chain-of-Thought Prompting in Large Language Models via Gradient-based Feature Attributions
Figure 4 for Analyzing Chain-of-Thought Prompting in Large Language Models via Gradient-based Feature Attributions
Viaarxiv icon

Consistent Explanations in the Face of Model Indeterminacy via Ensembling

Add code
Bookmark button
Alert button
Jun 13, 2023
Dan Ley, Leonard Tang, Matthew Nazari, Hongjin Lin, Suraj Srinivas, Himabindu Lakkaraju

Figure 1 for Consistent Explanations in the Face of Model Indeterminacy via Ensembling
Figure 2 for Consistent Explanations in the Face of Model Indeterminacy via Ensembling
Figure 3 for Consistent Explanations in the Face of Model Indeterminacy via Ensembling
Figure 4 for Consistent Explanations in the Face of Model Indeterminacy via Ensembling
Viaarxiv icon

On Minimizing the Impact of Dataset Shifts on Actionable Explanations

Add code
Bookmark button
Alert button
Jun 11, 2023
Anna P. Meyer, Dan Ley, Suraj Srinivas, Himabindu Lakkaraju

Figure 1 for On Minimizing the Impact of Dataset Shifts on Actionable Explanations
Figure 2 for On Minimizing the Impact of Dataset Shifts on Actionable Explanations
Figure 3 for On Minimizing the Impact of Dataset Shifts on Actionable Explanations
Figure 4 for On Minimizing the Impact of Dataset Shifts on Actionable Explanations
Viaarxiv icon