Picture for Preethi Lahoti

Preethi Lahoti

Inducing Group Fairness in LLM-Based Decisions

Add code
Jun 24, 2024
Viaarxiv icon

Automated Adversarial Discovery for Safety Classifiers

Add code
Jun 24, 2024
Viaarxiv icon

Gemini: A Family of Highly Capable Multimodal Models

Add code
Dec 19, 2023
Viaarxiv icon

FRAPPÉ: A Post-Processing Framework for Group Fairness Regularization

Add code
Dec 05, 2023
Figure 1 for FRAPPÉ: A Post-Processing Framework for Group Fairness Regularization
Figure 2 for FRAPPÉ: A Post-Processing Framework for Group Fairness Regularization
Figure 3 for FRAPPÉ: A Post-Processing Framework for Group Fairness Regularization
Figure 4 for FRAPPÉ: A Post-Processing Framework for Group Fairness Regularization
Viaarxiv icon

AART: AI-Assisted Red-Teaming with Diverse Data Generation for New LLM-powered Applications

Add code
Nov 29, 2023
Figure 1 for AART: AI-Assisted Red-Teaming with Diverse Data Generation for New LLM-powered Applications
Figure 2 for AART: AI-Assisted Red-Teaming with Diverse Data Generation for New LLM-powered Applications
Figure 3 for AART: AI-Assisted Red-Teaming with Diverse Data Generation for New LLM-powered Applications
Figure 4 for AART: AI-Assisted Red-Teaming with Diverse Data Generation for New LLM-powered Applications
Viaarxiv icon

Improving Diversity of Demographic Representation in Large Language Models via Collective-Critiques and Self-Voting

Add code
Oct 25, 2023
Figure 1 for Improving Diversity of Demographic Representation in Large Language Models via Collective-Critiques and Self-Voting
Figure 2 for Improving Diversity of Demographic Representation in Large Language Models via Collective-Critiques and Self-Voting
Figure 3 for Improving Diversity of Demographic Representation in Large Language Models via Collective-Critiques and Self-Voting
Figure 4 for Improving Diversity of Demographic Representation in Large Language Models via Collective-Critiques and Self-Voting
Viaarxiv icon

Detecting and Mitigating Test-time Failure Risks via Model-agnostic Uncertainty Learning

Add code
Sep 09, 2021
Figure 1 for Detecting and Mitigating Test-time Failure Risks via Model-agnostic Uncertainty Learning
Figure 2 for Detecting and Mitigating Test-time Failure Risks via Model-agnostic Uncertainty Learning
Figure 3 for Detecting and Mitigating Test-time Failure Risks via Model-agnostic Uncertainty Learning
Figure 4 for Detecting and Mitigating Test-time Failure Risks via Model-agnostic Uncertainty Learning
Viaarxiv icon

Accounting for Model Uncertainty in Algorithmic Discrimination

Add code
May 10, 2021
Figure 1 for Accounting for Model Uncertainty in Algorithmic Discrimination
Figure 2 for Accounting for Model Uncertainty in Algorithmic Discrimination
Figure 3 for Accounting for Model Uncertainty in Algorithmic Discrimination
Figure 4 for Accounting for Model Uncertainty in Algorithmic Discrimination
Viaarxiv icon

Fairness without Demographics through Adversarially Reweighted Learning

Add code
Jun 24, 2020
Figure 1 for Fairness without Demographics through Adversarially Reweighted Learning
Figure 2 for Fairness without Demographics through Adversarially Reweighted Learning
Figure 3 for Fairness without Demographics through Adversarially Reweighted Learning
Figure 4 for Fairness without Demographics through Adversarially Reweighted Learning
Viaarxiv icon

An Empirical Study on Learning Fairness Metrics for COMPAS Data with Human Supervision

Add code
Oct 31, 2019
Figure 1 for An Empirical Study on Learning Fairness Metrics for COMPAS Data with Human Supervision
Figure 2 for An Empirical Study on Learning Fairness Metrics for COMPAS Data with Human Supervision
Figure 3 for An Empirical Study on Learning Fairness Metrics for COMPAS Data with Human Supervision
Viaarxiv icon