Picture for Akari Asai

Akari Asai

CopyBench: Measuring Literal and Non-Literal Reproduction of Copyright-Protected Text in Language Model Generation

Add code
Jul 09, 2024
Viaarxiv icon

CodeRAG-Bench: Can Retrieval Augment Code Generation?

Add code
Jun 20, 2024
Figure 1 for CodeRAG-Bench: Can Retrieval Augment Code Generation?
Figure 2 for CodeRAG-Bench: Can Retrieval Augment Code Generation?
Figure 3 for CodeRAG-Bench: Can Retrieval Augment Code Generation?
Figure 4 for CodeRAG-Bench: Can Retrieval Augment Code Generation?
Viaarxiv icon

Reliable, Adaptable, and Attributable Language Models with Retrieval

Add code
Mar 05, 2024
Figure 1 for Reliable, Adaptable, and Attributable Language Models with Retrieval
Figure 2 for Reliable, Adaptable, and Attributable Language Models with Retrieval
Figure 3 for Reliable, Adaptable, and Attributable Language Models with Retrieval
Figure 4 for Reliable, Adaptable, and Attributable Language Models with Retrieval
Viaarxiv icon

Fine-grained Hallucination Detection and Editing for Language Models

Add code
Jan 17, 2024
Viaarxiv icon

Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection

Add code
Oct 17, 2023
Figure 1 for Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection
Figure 2 for Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection
Figure 3 for Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection
Figure 4 for Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection
Viaarxiv icon

BUFFET: Benchmarking Large Language Models for Few-shot Cross-lingual Transfer

Add code
May 24, 2023
Figure 1 for BUFFET: Benchmarking Large Language Models for Few-shot Cross-lingual Transfer
Figure 2 for BUFFET: Benchmarking Large Language Models for Few-shot Cross-lingual Transfer
Figure 3 for BUFFET: Benchmarking Large Language Models for Few-shot Cross-lingual Transfer
Figure 4 for BUFFET: Benchmarking Large Language Models for Few-shot Cross-lingual Transfer
Viaarxiv icon

TaskWeb: Selecting Better Source Tasks for Multi-task NLP

Add code
May 22, 2023
Figure 1 for TaskWeb: Selecting Better Source Tasks for Multi-task NLP
Figure 2 for TaskWeb: Selecting Better Source Tasks for Multi-task NLP
Figure 3 for TaskWeb: Selecting Better Source Tasks for Multi-task NLP
Figure 4 for TaskWeb: Selecting Better Source Tasks for Multi-task NLP
Viaarxiv icon

xPQA: Cross-Lingual Product Question Answering across 12 Languages

Add code
May 16, 2023
Figure 1 for xPQA: Cross-Lingual Product Question Answering across 12 Languages
Figure 2 for xPQA: Cross-Lingual Product Question Answering across 12 Languages
Figure 3 for xPQA: Cross-Lingual Product Question Answering across 12 Languages
Figure 4 for xPQA: Cross-Lingual Product Question Answering across 12 Languages
Viaarxiv icon

AfriQA: Cross-lingual Open-Retrieval Question Answering for African Languages

Add code
May 11, 2023
Figure 1 for AfriQA: Cross-lingual Open-Retrieval Question Answering for African Languages
Figure 2 for AfriQA: Cross-lingual Open-Retrieval Question Answering for African Languages
Figure 3 for AfriQA: Cross-lingual Open-Retrieval Question Answering for African Languages
Figure 4 for AfriQA: Cross-lingual Open-Retrieval Question Answering for African Languages
Viaarxiv icon

How to Train Your DRAGON: Diverse Augmentation Towards Generalizable Dense Retrieval

Add code
Feb 15, 2023
Figure 1 for How to Train Your DRAGON: Diverse Augmentation Towards Generalizable Dense Retrieval
Figure 2 for How to Train Your DRAGON: Diverse Augmentation Towards Generalizable Dense Retrieval
Figure 3 for How to Train Your DRAGON: Diverse Augmentation Towards Generalizable Dense Retrieval
Figure 4 for How to Train Your DRAGON: Diverse Augmentation Towards Generalizable Dense Retrieval
Viaarxiv icon