Picture for Nouha Dziri

Nouha Dziri

WildBench: Benchmarking LLMs with Challenging Tasks from Real Users in the Wild

Add code
Jun 07, 2024
Viaarxiv icon

CULTURE-GEN: Revealing Global Cultural Perception in Language Models through Natural Language Prompting

Add code
Apr 16, 2024
Figure 1 for CULTURE-GEN: Revealing Global Cultural Perception in Language Models through Natural Language Prompting
Figure 2 for CULTURE-GEN: Revealing Global Cultural Perception in Language Models through Natural Language Prompting
Figure 3 for CULTURE-GEN: Revealing Global Cultural Perception in Language Models through Natural Language Prompting
Figure 4 for CULTURE-GEN: Revealing Global Cultural Perception in Language Models through Natural Language Prompting
Viaarxiv icon

RewardBench: Evaluating Reward Models for Language Modeling

Add code
Mar 20, 2024
Figure 1 for RewardBench: Evaluating Reward Models for Language Modeling
Figure 2 for RewardBench: Evaluating Reward Models for Language Modeling
Figure 3 for RewardBench: Evaluating Reward Models for Language Modeling
Figure 4 for RewardBench: Evaluating Reward Models for Language Modeling
Viaarxiv icon

A Roadmap to Pluralistic Alignment

Add code
Feb 07, 2024
Viaarxiv icon

The Unlocking Spell on Base LLMs: Rethinking Alignment via In-Context Learning

Add code
Dec 04, 2023
Figure 1 for The Unlocking Spell on Base LLMs: Rethinking Alignment via In-Context Learning
Figure 2 for The Unlocking Spell on Base LLMs: Rethinking Alignment via In-Context Learning
Figure 3 for The Unlocking Spell on Base LLMs: Rethinking Alignment via In-Context Learning
Figure 4 for The Unlocking Spell on Base LLMs: Rethinking Alignment via In-Context Learning
Viaarxiv icon

What Makes it Ok to Set a Fire? Iterative Self-distillation of Contexts and Rationales for Disambiguating Defeasible Social and Moral Situations

Nov 01, 2023
Figure 1 for What Makes it Ok to Set a Fire? Iterative Self-distillation of Contexts and Rationales for Disambiguating Defeasible Social and Moral Situations
Figure 2 for What Makes it Ok to Set a Fire? Iterative Self-distillation of Contexts and Rationales for Disambiguating Defeasible Social and Moral Situations
Figure 3 for What Makes it Ok to Set a Fire? Iterative Self-distillation of Contexts and Rationales for Disambiguating Defeasible Social and Moral Situations
Figure 4 for What Makes it Ok to Set a Fire? Iterative Self-distillation of Contexts and Rationales for Disambiguating Defeasible Social and Moral Situations
Viaarxiv icon

The Generative AI Paradox: "What It Can Create, It May Not Understand"

Oct 31, 2023
Figure 1 for The Generative AI Paradox: "What It Can Create, It May Not Understand"
Figure 2 for The Generative AI Paradox: "What It Can Create, It May Not Understand"
Figure 3 for The Generative AI Paradox: "What It Can Create, It May Not Understand"
Figure 4 for The Generative AI Paradox: "What It Can Create, It May Not Understand"
Viaarxiv icon

Phenomenal Yet Puzzling: Testing Inductive Reasoning Capabilities of Language Models with Hypothesis Refinement

Add code
Oct 12, 2023
Figure 1 for Phenomenal Yet Puzzling: Testing Inductive Reasoning Capabilities of Language Models with Hypothesis Refinement
Figure 2 for Phenomenal Yet Puzzling: Testing Inductive Reasoning Capabilities of Language Models with Hypothesis Refinement
Figure 3 for Phenomenal Yet Puzzling: Testing Inductive Reasoning Capabilities of Language Models with Hypothesis Refinement
Figure 4 for Phenomenal Yet Puzzling: Testing Inductive Reasoning Capabilities of Language Models with Hypothesis Refinement
Viaarxiv icon

Value Kaleidoscope: Engaging AI with Pluralistic Human Values, Rights, and Duties

Add code
Sep 02, 2023
Figure 1 for Value Kaleidoscope: Engaging AI with Pluralistic Human Values, Rights, and Duties
Figure 2 for Value Kaleidoscope: Engaging AI with Pluralistic Human Values, Rights, and Duties
Figure 3 for Value Kaleidoscope: Engaging AI with Pluralistic Human Values, Rights, and Duties
Figure 4 for Value Kaleidoscope: Engaging AI with Pluralistic Human Values, Rights, and Duties
Viaarxiv icon

Fine-Grained Human Feedback Gives Better Rewards for Language Model Training

Add code
Jun 02, 2023
Figure 1 for Fine-Grained Human Feedback Gives Better Rewards for Language Model Training
Figure 2 for Fine-Grained Human Feedback Gives Better Rewards for Language Model Training
Figure 3 for Fine-Grained Human Feedback Gives Better Rewards for Language Model Training
Figure 4 for Fine-Grained Human Feedback Gives Better Rewards for Language Model Training
Viaarxiv icon