Picture for Allyson Ettinger

Allyson Ettinger

FlexOlmo: Open Language Models for Flexible Data Use

Add code
Jul 09, 2025
Viaarxiv icon

To Err is AI : A Case Study Informing LLM Flaw Reporting Practices

Add code
Oct 15, 2024
Figure 1 for To Err is AI : A Case Study Informing LLM Flaw Reporting Practices
Figure 2 for To Err is AI : A Case Study Informing LLM Flaw Reporting Practices
Figure 3 for To Err is AI : A Case Study Informing LLM Flaw Reporting Practices
Figure 4 for To Err is AI : A Case Study Informing LLM Flaw Reporting Practices
Viaarxiv icon

AI as Humanity's Salieri: Quantifying Linguistic Creativity of Language Models via Systematic Attribution of Machine Text against Web Text

Add code
Oct 05, 2024
Viaarxiv icon

WildGuard: Open One-Stop Moderation Tools for Safety Risks, Jailbreaks, and Refusals of LLMs

Add code
Jun 26, 2024
Figure 1 for WildGuard: Open One-Stop Moderation Tools for Safety Risks, Jailbreaks, and Refusals of LLMs
Figure 2 for WildGuard: Open One-Stop Moderation Tools for Safety Risks, Jailbreaks, and Refusals of LLMs
Figure 3 for WildGuard: Open One-Stop Moderation Tools for Safety Risks, Jailbreaks, and Refusals of LLMs
Figure 4 for WildGuard: Open One-Stop Moderation Tools for Safety Risks, Jailbreaks, and Refusals of LLMs
Viaarxiv icon

WildTeaming at Scale: From In-the-Wild Jailbreaks to (Adversarially) Safer Language Models

Add code
Jun 26, 2024
Figure 1 for WildTeaming at Scale: From In-the-Wild Jailbreaks to (Adversarially) Safer Language Models
Figure 2 for WildTeaming at Scale: From In-the-Wild Jailbreaks to (Adversarially) Safer Language Models
Figure 3 for WildTeaming at Scale: From In-the-Wild Jailbreaks to (Adversarially) Safer Language Models
Figure 4 for WildTeaming at Scale: From In-the-Wild Jailbreaks to (Adversarially) Safer Language Models
Viaarxiv icon

When Hindsight is Not 20/20: Testing Limits on Reflective Thinking in Large Language Models

Add code
Apr 14, 2024
Figure 1 for When Hindsight is Not 20/20: Testing Limits on Reflective Thinking in Large Language Models
Figure 2 for When Hindsight is Not 20/20: Testing Limits on Reflective Thinking in Large Language Models
Figure 3 for When Hindsight is Not 20/20: Testing Limits on Reflective Thinking in Large Language Models
Figure 4 for When Hindsight is Not 20/20: Testing Limits on Reflective Thinking in Large Language Models
Viaarxiv icon

Experimental Contexts Can Facilitate Robust Semantic Property Inference in Language Models, but Inconsistently

Add code
Jan 12, 2024
Figure 1 for Experimental Contexts Can Facilitate Robust Semantic Property Inference in Language Models, but Inconsistently
Figure 2 for Experimental Contexts Can Facilitate Robust Semantic Property Inference in Language Models, but Inconsistently
Figure 3 for Experimental Contexts Can Facilitate Robust Semantic Property Inference in Language Models, but Inconsistently
Figure 4 for Experimental Contexts Can Facilitate Robust Semantic Property Inference in Language Models, but Inconsistently
Viaarxiv icon

The Generative AI Paradox: "What It Can Create, It May Not Understand"

Add code
Oct 31, 2023
Viaarxiv icon

"You Are An Expert Linguistic Annotator": Limits of LLMs as Analyzers of Abstract Meaning Representation

Add code
Oct 26, 2023
Figure 1 for "You Are An Expert Linguistic Annotator": Limits of LLMs as Analyzers of Abstract Meaning Representation
Figure 2 for "You Are An Expert Linguistic Annotator": Limits of LLMs as Analyzers of Abstract Meaning Representation
Figure 3 for "You Are An Expert Linguistic Annotator": Limits of LLMs as Analyzers of Abstract Meaning Representation
Figure 4 for "You Are An Expert Linguistic Annotator": Limits of LLMs as Analyzers of Abstract Meaning Representation
Viaarxiv icon

Can You Follow Me? Testing Situational Understanding in ChatGPT

Add code
Oct 24, 2023
Figure 1 for Can You Follow Me? Testing Situational Understanding in ChatGPT
Figure 2 for Can You Follow Me? Testing Situational Understanding in ChatGPT
Figure 3 for Can You Follow Me? Testing Situational Understanding in ChatGPT
Figure 4 for Can You Follow Me? Testing Situational Understanding in ChatGPT
Viaarxiv icon