Alert button
Picture for Adina Williams

Adina Williams

Alert button

Introducing v0.5 of the AI Safety Benchmark from MLCommons

Add code
Bookmark button
Alert button
Apr 18, 2024
Bertie Vidgen, Adarsh Agrawal, Ahmed M. Ahmed, Victor Akinwande, Namir Al-Nuaimi, Najla Alfaraj, Elie Alhajjar, Lora Aroyo, Trupti Bavalatti, Borhane Blili-Hamelin, Kurt Bollacker, Rishi Bomassani, Marisa Ferrara Boston, Siméon Campos, Kal Chakra, Canyu Chen, Cody Coleman, Zacharie Delpierre Coudert, Leon Derczynski, Debojyoti Dutta, Ian Eisenberg, James Ezick, Heather Frase, Brian Fuller, Ram Gandikota, Agasthya Gangavarapu, Ananya Gangavarapu, James Gealy, Rajat Ghosh, James Goel, Usman Gohar, Sujata Goswami, Scott A. Hale, Wiebke Hutiri, Joseph Marvin Imperial, Surgan Jandial, Nick Judd, Felix Juefei-Xu, Foutse Khomh, Bhavya Kailkhura, Hannah Rose Kirk, Kevin Klyman, Chris Knotz, Michael Kuchnik, Shachi H. Kumar, Chris Lengerich, Bo Li, Zeyi Liao, Eileen Peters Long, Victor Lu, Yifan Mai, Priyanka Mary Mammen, Kelvin Manyeki, Sean McGregor, Virendra Mehta, Shafee Mohammed, Emanuel Moss, Lama Nachman, Dinesh Jinenhally Naganna, Amin Nikanjam, Besmira Nushi, Luis Oala, Iftach Orr, Alicia Parrish, Cigdem Patlak, William Pietri, Forough Poursabzi-Sangdeh, Eleonora Presani, Fabrizio Puletti, Paul Röttger, Saurav Sahay, Tim Santos, Nino Scherrer, Alice Schoenauer Sebag, Patrick Schramowski, Abolfazl Shahbazi, Vin Sharma, Xudong Shen, Vamsi Sistla, Leonard Tang, Davide Testuggine, Vithursan Thangarasa, Elizabeth Anne Watkins, Rebecca Weiss, Chris Welty, Tyler Wilbers, Adina Williams, Carole-Jean Wu, Poonam Yadav, Xianjun Yang, Yi Zeng, Wenhui Zhang, Fedor Zhdanov, Jiacheng Zhu, Percy Liang, Peter Mattson, Joaquin Vanschoren

Viaarxiv icon

[Call for Papers] The 2nd BabyLM Challenge: Sample-efficient pretraining on a developmentally plausible corpus

Add code
Bookmark button
Alert button
Apr 09, 2024
Leshem Choshen, Ryan Cotterell, Michael Y. Hu, Tal Linzen, Aaron Mueller, Candace Ross, Alex Warstadt, Ethan Wilcox, Adina Williams, Chengxu Zhuang

Viaarxiv icon

Improving Text-to-Image Consistency via Automatic Prompt Optimization

Add code
Bookmark button
Alert button
Mar 26, 2024
Oscar Mañas, Pietro Astolfi, Melissa Hall, Candace Ross, Jack Urbanek, Adina Williams, Aishwarya Agrawal, Adriana Romero-Soriano, Michal Drozdzal

Viaarxiv icon

Compositional learning of functions in humans and machines

Add code
Bookmark button
Alert button
Mar 18, 2024
Yanli Zhou, Brenden M. Lake, Adina Williams

Figure 1 for Compositional learning of functions in humans and machines
Figure 2 for Compositional learning of functions in humans and machines
Figure 3 for Compositional learning of functions in humans and machines
Figure 4 for Compositional learning of functions in humans and machines
Viaarxiv icon

EmphAssess : a Prosodic Benchmark on Assessing Emphasis Transfer in Speech-to-Speech Models

Add code
Bookmark button
Alert button
Dec 21, 2023
Maureen de Seyssel, Antony D'Avirro, Adina Williams, Emmanuel Dupoux

Viaarxiv icon

Grammatical Gender's Influence on Distributional Semantics: A Causal Perspective

Add code
Bookmark button
Alert button
Nov 30, 2023
Karolina Stańczak, Kevin Du, Adina Williams, Isabelle Augenstein, Ryan Cotterell

Figure 1 for Grammatical Gender's Influence on Distributional Semantics: A Causal Perspective
Figure 2 for Grammatical Gender's Influence on Distributional Semantics: A Causal Perspective
Figure 3 for Grammatical Gender's Influence on Distributional Semantics: A Causal Perspective
Figure 4 for Grammatical Gender's Influence on Distributional Semantics: A Causal Perspective
Viaarxiv icon

ROBBIE: Robust Bias Evaluation of Large Generative Language Models

Add code
Bookmark button
Alert button
Nov 29, 2023
David Esiobu, Xiaoqing Tan, Saghar Hosseini, Megan Ung, Yuchen Zhang, Jude Fernandes, Jane Dwivedi-Yu, Eleonora Presani, Adina Williams, Eric Michael Smith

Figure 1 for ROBBIE: Robust Bias Evaluation of Large Generative Language Models
Figure 2 for ROBBIE: Robust Bias Evaluation of Large Generative Language Models
Figure 3 for ROBBIE: Robust Bias Evaluation of Large Generative Language Models
Figure 4 for ROBBIE: Robust Bias Evaluation of Large Generative Language Models
Viaarxiv icon

The Validity of Evaluation Results: Assessing Concurrence Across Compositionality Benchmarks

Add code
Bookmark button
Alert button
Oct 26, 2023
Kaiser Sun, Adina Williams, Dieuwke Hupkes

Viaarxiv icon

The Gender-GAP Pipeline: A Gender-Aware Polyglot Pipeline for Gender Characterisation in 55 Languages

Add code
Bookmark button
Alert button
Aug 31, 2023
Benjamin Muller, Belen Alastruey, Prangthip Hansanti, Elahe Kalbassi, Christophe Ropers, Eric Michael Smith, Adina Williams, Luke Zettlemoyer, Pierre Andrews, Marta R. Costa-jussà

Figure 1 for The Gender-GAP Pipeline: A Gender-Aware Polyglot Pipeline for Gender Characterisation in 55 Languages
Figure 2 for The Gender-GAP Pipeline: A Gender-Aware Polyglot Pipeline for Gender Characterisation in 55 Languages
Figure 3 for The Gender-GAP Pipeline: A Gender-Aware Polyglot Pipeline for Gender Characterisation in 55 Languages
Figure 4 for The Gender-GAP Pipeline: A Gender-Aware Polyglot Pipeline for Gender Characterisation in 55 Languages
Viaarxiv icon

DIG In: Evaluating Disparities in Image Generations with Indicators for Geographic Diversity

Add code
Bookmark button
Alert button
Aug 15, 2023
Melissa Hall, Candace Ross, Adina Williams, Nicolas Carion, Michal Drozdzal, Adriana Romero Soriano

Figure 1 for DIG In: Evaluating Disparities in Image Generations with Indicators for Geographic Diversity
Figure 2 for DIG In: Evaluating Disparities in Image Generations with Indicators for Geographic Diversity
Figure 3 for DIG In: Evaluating Disparities in Image Generations with Indicators for Geographic Diversity
Figure 4 for DIG In: Evaluating Disparities in Image Generations with Indicators for Geographic Diversity
Viaarxiv icon