Picture for Khyathi Chandu

Khyathi Chandu

DataComp-LM: In search of the next generation of training sets for language models

Add code
Jun 18, 2024
Viaarxiv icon

WildBench: Benchmarking LLMs with Challenging Tasks from Real Users in the Wild

Add code
Jun 07, 2024
Viaarxiv icon

On the Role of Summary Content Units in Text Summarization Evaluation

Add code
Apr 02, 2024
Figure 1 for On the Role of Summary Content Units in Text Summarization Evaluation
Figure 2 for On the Role of Summary Content Units in Text Summarization Evaluation
Figure 3 for On the Role of Summary Content Units in Text Summarization Evaluation
Figure 4 for On the Role of Summary Content Units in Text Summarization Evaluation
Viaarxiv icon

RewardBench: Evaluating Reward Models for Language Modeling

Add code
Mar 20, 2024
Figure 1 for RewardBench: Evaluating Reward Models for Language Modeling
Figure 2 for RewardBench: Evaluating Reward Models for Language Modeling
Figure 3 for RewardBench: Evaluating Reward Models for Language Modeling
Figure 4 for RewardBench: Evaluating Reward Models for Language Modeling
Viaarxiv icon

L3GO: Language Agents with Chain-of-3D-Thoughts for Generating Unconventional Objects

Add code
Feb 14, 2024
Figure 1 for L3GO: Language Agents with Chain-of-3D-Thoughts for Generating Unconventional Objects
Figure 2 for L3GO: Language Agents with Chain-of-3D-Thoughts for Generating Unconventional Objects
Figure 3 for L3GO: Language Agents with Chain-of-3D-Thoughts for Generating Unconventional Objects
Figure 4 for L3GO: Language Agents with Chain-of-3D-Thoughts for Generating Unconventional Objects
Viaarxiv icon

Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research

Add code
Jan 31, 2024
Figure 1 for Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research
Figure 2 for Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research
Figure 3 for Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research
Figure 4 for Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research
Viaarxiv icon

NovaCOMET: Open Commonsense Foundation Models with Symbolic Knowledge Distillation

Add code
Dec 10, 2023
Figure 1 for NovaCOMET: Open Commonsense Foundation Models with Symbolic Knowledge Distillation
Figure 2 for NovaCOMET: Open Commonsense Foundation Models with Symbolic Knowledge Distillation
Figure 3 for NovaCOMET: Open Commonsense Foundation Models with Symbolic Knowledge Distillation
Figure 4 for NovaCOMET: Open Commonsense Foundation Models with Symbolic Knowledge Distillation
Viaarxiv icon

The Unlocking Spell on Base LLMs: Rethinking Alignment via In-Context Learning

Add code
Dec 04, 2023
Figure 1 for The Unlocking Spell on Base LLMs: Rethinking Alignment via In-Context Learning
Figure 2 for The Unlocking Spell on Base LLMs: Rethinking Alignment via In-Context Learning
Figure 3 for The Unlocking Spell on Base LLMs: Rethinking Alignment via In-Context Learning
Figure 4 for The Unlocking Spell on Base LLMs: Rethinking Alignment via In-Context Learning
Viaarxiv icon

Lumos: Learning Agents with Unified Data, Modular Design, and Open-Source LLMs

Add code
Nov 09, 2023
Figure 1 for Lumos: Learning Agents with Unified Data, Modular Design, and Open-Source LLMs
Figure 2 for Lumos: Learning Agents with Unified Data, Modular Design, and Open-Source LLMs
Figure 3 for Lumos: Learning Agents with Unified Data, Modular Design, and Open-Source LLMs
Figure 4 for Lumos: Learning Agents with Unified Data, Modular Design, and Open-Source LLMs
Viaarxiv icon

The Generative AI Paradox: "What It Can Create, It May Not Understand"

Add code
Oct 31, 2023
Figure 1 for The Generative AI Paradox: "What It Can Create, It May Not Understand"
Figure 2 for The Generative AI Paradox: "What It Can Create, It May Not Understand"
Figure 3 for The Generative AI Paradox: "What It Can Create, It May Not Understand"
Figure 4 for The Generative AI Paradox: "What It Can Create, It May Not Understand"
Viaarxiv icon