Picture for Ioana Baldini

Ioana Baldini

Why Don't Prompt-Based Fairness Metrics Correlate?

Add code
Jun 09, 2024
Viaarxiv icon

Detectors for Safe and Reliable LLMs: Implementations, Uses, and Limitations

Add code
Mar 09, 2024
Figure 1 for Detectors for Safe and Reliable LLMs: Implementations, Uses, and Limitations
Figure 2 for Detectors for Safe and Reliable LLMs: Implementations, Uses, and Limitations
Figure 3 for Detectors for Safe and Reliable LLMs: Implementations, Uses, and Limitations
Figure 4 for Detectors for Safe and Reliable LLMs: Implementations, Uses, and Limitations
Viaarxiv icon

Alignment Studio: Aligning Large Language Models to Particular Contextual Regulations

Add code
Mar 08, 2024
Figure 1 for Alignment Studio: Aligning Large Language Models to Particular Contextual Regulations
Figure 2 for Alignment Studio: Aligning Large Language Models to Particular Contextual Regulations
Figure 3 for Alignment Studio: Aligning Large Language Models to Particular Contextual Regulations
Figure 4 for Alignment Studio: Aligning Large Language Models to Particular Contextual Regulations
Viaarxiv icon

SocialStigmaQA: A Benchmark to Uncover Stigma Amplification in Generative Language Models

Add code
Dec 27, 2023
Viaarxiv icon

Fairness-Aware Structured Pruning in Transformers

Add code
Dec 24, 2023
Figure 1 for Fairness-Aware Structured Pruning in Transformers
Figure 2 for Fairness-Aware Structured Pruning in Transformers
Figure 3 for Fairness-Aware Structured Pruning in Transformers
Figure 4 for Fairness-Aware Structured Pruning in Transformers
Viaarxiv icon

Subtle Misogyny Detection and Mitigation: An Expert-Annotated Dataset

Add code
Nov 15, 2023
Figure 1 for Subtle Misogyny Detection and Mitigation: An Expert-Annotated Dataset
Figure 2 for Subtle Misogyny Detection and Mitigation: An Expert-Annotated Dataset
Figure 3 for Subtle Misogyny Detection and Mitigation: An Expert-Annotated Dataset
Figure 4 for Subtle Misogyny Detection and Mitigation: An Expert-Annotated Dataset
Viaarxiv icon

Keeping Up with the Language Models: Robustness-Bias Interplay in NLI Data and Models

Add code
May 22, 2023
Figure 1 for Keeping Up with the Language Models: Robustness-Bias Interplay in NLI Data and Models
Figure 2 for Keeping Up with the Language Models: Robustness-Bias Interplay in NLI Data and Models
Figure 3 for Keeping Up with the Language Models: Robustness-Bias Interplay in NLI Data and Models
Figure 4 for Keeping Up with the Language Models: Robustness-Bias Interplay in NLI Data and Models
Viaarxiv icon

Write It Like You See It: Detectable Differences in Clinical Notes By Race Lead To Differential Model Recommendations

Add code
May 08, 2022
Figure 1 for Write It Like You See It: Detectable Differences in Clinical Notes By Race Lead To Differential Model Recommendations
Figure 2 for Write It Like You See It: Detectable Differences in Clinical Notes By Race Lead To Differential Model Recommendations
Figure 3 for Write It Like You See It: Detectable Differences in Clinical Notes By Race Lead To Differential Model Recommendations
Figure 4 for Write It Like You See It: Detectable Differences in Clinical Notes By Race Lead To Differential Model Recommendations
Viaarxiv icon

Downstream Fairness Caveats with Synthetic Healthcare Data

Add code
Mar 09, 2022
Figure 1 for Downstream Fairness Caveats with Synthetic Healthcare Data
Figure 2 for Downstream Fairness Caveats with Synthetic Healthcare Data
Figure 3 for Downstream Fairness Caveats with Synthetic Healthcare Data
Figure 4 for Downstream Fairness Caveats with Synthetic Healthcare Data
Viaarxiv icon

Ground-Truth, Whose Truth? -- Examining the Challenges with Annotating Toxic Text Datasets

Add code
Dec 07, 2021
Figure 1 for Ground-Truth, Whose Truth? -- Examining the Challenges with Annotating Toxic Text Datasets
Figure 2 for Ground-Truth, Whose Truth? -- Examining the Challenges with Annotating Toxic Text Datasets
Figure 3 for Ground-Truth, Whose Truth? -- Examining the Challenges with Annotating Toxic Text Datasets
Figure 4 for Ground-Truth, Whose Truth? -- Examining the Challenges with Annotating Toxic Text Datasets
Viaarxiv icon