Alert button
Picture for Ioana Baldini

Ioana Baldini

Alert button

Detectors for Safe and Reliable LLMs: Implementations, Uses, and Limitations

Add code
Bookmark button
Alert button
Mar 09, 2024
Swapnaja Achintalwar, Adriana Alvarado Garcia, Ateret Anaby-Tavor, Ioana Baldini, Sara E. Berger, Bishwaranjan Bhattacharjee, Djallel Bouneffouf, Subhajit Chaudhury, Pin-Yu Chen, Lamogha Chiazor, Elizabeth M. Daly, Rogério Abreu de Paula, Pierre Dognin, Eitan Farchi, Soumya Ghosh, Michael Hind, Raya Horesh, George Kour, Ja Young Lee, Erik Miehling, Keerthiram Murugesan, Manish Nagireddy, Inkit Padhi, David Piorkowski, Ambrish Rawat, Orna Raz, Prasanna Sattigeri, Hendrik Strobelt, Sarathkrishna Swaminathan, Christoph Tillmann, Aashka Trivedi, Kush R. Varshney, Dennis Wei, Shalisha Witherspooon, Marcel Zalmanovici

Figure 1 for Detectors for Safe and Reliable LLMs: Implementations, Uses, and Limitations
Figure 2 for Detectors for Safe and Reliable LLMs: Implementations, Uses, and Limitations
Figure 3 for Detectors for Safe and Reliable LLMs: Implementations, Uses, and Limitations
Figure 4 for Detectors for Safe and Reliable LLMs: Implementations, Uses, and Limitations
Viaarxiv icon

Alignment Studio: Aligning Large Language Models to Particular Contextual Regulations

Add code
Bookmark button
Alert button
Mar 08, 2024
Swapnaja Achintalwar, Ioana Baldini, Djallel Bouneffouf, Joan Byamugisha, Maria Chang, Pierre Dognin, Eitan Farchi, Ndivhuwo Makondo, Aleksandra Mojsilovic, Manish Nagireddy, Karthikeyan Natesan Ramamurthy, Inkit Padhi, Orna Raz, Jesus Rios, Prasanna Sattigeri, Moninder Singh, Siphiwe Thwala, Rosario A. Uceda-Sosa, Kush R. Varshney

Figure 1 for Alignment Studio: Aligning Large Language Models to Particular Contextual Regulations
Figure 2 for Alignment Studio: Aligning Large Language Models to Particular Contextual Regulations
Figure 3 for Alignment Studio: Aligning Large Language Models to Particular Contextual Regulations
Figure 4 for Alignment Studio: Aligning Large Language Models to Particular Contextual Regulations
Viaarxiv icon

SocialStigmaQA: A Benchmark to Uncover Stigma Amplification in Generative Language Models

Add code
Bookmark button
Alert button
Dec 27, 2023
Manish Nagireddy, Lamogha Chiazor, Moninder Singh, Ioana Baldini

Viaarxiv icon

Fairness-Aware Structured Pruning in Transformers

Add code
Bookmark button
Alert button
Dec 24, 2023
Abdelrahman Zayed, Goncalo Mordido, Samira Shabanian, Ioana Baldini, Sarath Chandar

Viaarxiv icon

Subtle Misogyny Detection and Mitigation: An Expert-Annotated Dataset

Add code
Bookmark button
Alert button
Nov 15, 2023
Brooklyn Sheppard, Anna Richter, Allison Cohen, Elizabeth Allyn Smith, Tamara Kneese, Carolyne Pelletier, Ioana Baldini, Yue Dong

Viaarxiv icon

Keeping Up with the Language Models: Robustness-Bias Interplay in NLI Data and Models

Add code
Bookmark button
Alert button
May 22, 2023
Ioana Baldini, Chhavi Yadav, Payel Das, Kush R. Varshney

Figure 1 for Keeping Up with the Language Models: Robustness-Bias Interplay in NLI Data and Models
Figure 2 for Keeping Up with the Language Models: Robustness-Bias Interplay in NLI Data and Models
Figure 3 for Keeping Up with the Language Models: Robustness-Bias Interplay in NLI Data and Models
Figure 4 for Keeping Up with the Language Models: Robustness-Bias Interplay in NLI Data and Models
Viaarxiv icon

Write It Like You See It: Detectable Differences in Clinical Notes By Race Lead To Differential Model Recommendations

Add code
Bookmark button
Alert button
May 08, 2022
Hammaad Adam, Ming Ying Yang, Kenrick Cato, Ioana Baldini, Charles Senteio, Leo Anthony Celi, Jiaming Zeng, Moninder Singh, Marzyeh Ghassemi

Figure 1 for Write It Like You See It: Detectable Differences in Clinical Notes By Race Lead To Differential Model Recommendations
Figure 2 for Write It Like You See It: Detectable Differences in Clinical Notes By Race Lead To Differential Model Recommendations
Figure 3 for Write It Like You See It: Detectable Differences in Clinical Notes By Race Lead To Differential Model Recommendations
Figure 4 for Write It Like You See It: Detectable Differences in Clinical Notes By Race Lead To Differential Model Recommendations
Viaarxiv icon

Downstream Fairness Caveats with Synthetic Healthcare Data

Add code
Bookmark button
Alert button
Mar 09, 2022
Karan Bhanot, Ioana Baldini, Dennis Wei, Jiaming Zeng, Kristin P. Bennett

Figure 1 for Downstream Fairness Caveats with Synthetic Healthcare Data
Figure 2 for Downstream Fairness Caveats with Synthetic Healthcare Data
Figure 3 for Downstream Fairness Caveats with Synthetic Healthcare Data
Figure 4 for Downstream Fairness Caveats with Synthetic Healthcare Data
Viaarxiv icon

Ground-Truth, Whose Truth? -- Examining the Challenges with Annotating Toxic Text Datasets

Add code
Bookmark button
Alert button
Dec 07, 2021
Kofi Arhin, Ioana Baldini, Dennis Wei, Karthikeyan Natesan Ramamurthy, Moninder Singh

Figure 1 for Ground-Truth, Whose Truth? -- Examining the Challenges with Annotating Toxic Text Datasets
Figure 2 for Ground-Truth, Whose Truth? -- Examining the Challenges with Annotating Toxic Text Datasets
Figure 3 for Ground-Truth, Whose Truth? -- Examining the Challenges with Annotating Toxic Text Datasets
Figure 4 for Ground-Truth, Whose Truth? -- Examining the Challenges with Annotating Toxic Text Datasets
Viaarxiv icon