Picture for Moontae Lee

Moontae Lee

Learning to Explore and Select for Coverage-Conditioned Retrieval-Augmented Generation

Add code
Jul 01, 2024
Viaarxiv icon

The BiGGen Bench: A Principled Benchmark for Fine-grained Evaluation of Language Models with Language Models

Add code
Jun 09, 2024
Viaarxiv icon

LG AI Research & KAIST at EHRSQL 2024: Self-Training Large Language Models with Pseudo-Labeled Unanswerable Questions for a Reliable Text-to-SQL System on EHRs

Add code
May 18, 2024
Viaarxiv icon

Understanding the Capabilities and Limitations of Large Language Models for Cultural Commonsense

Add code
May 07, 2024
Figure 1 for Understanding the Capabilities and Limitations of Large Language Models for Cultural Commonsense
Figure 2 for Understanding the Capabilities and Limitations of Large Language Models for Cultural Commonsense
Figure 3 for Understanding the Capabilities and Limitations of Large Language Models for Cultural Commonsense
Figure 4 for Understanding the Capabilities and Limitations of Large Language Models for Cultural Commonsense
Viaarxiv icon

Prometheus 2: An Open Source Language Model Specialized in Evaluating Other Language Models

Add code
May 02, 2024
Viaarxiv icon

Small Language Models Need Strong Verifiers to Self-Correct Reasoning

Add code
Apr 26, 2024
Figure 1 for Small Language Models Need Strong Verifiers to Self-Correct Reasoning
Figure 2 for Small Language Models Need Strong Verifiers to Self-Correct Reasoning
Figure 3 for Small Language Models Need Strong Verifiers to Self-Correct Reasoning
Figure 4 for Small Language Models Need Strong Verifiers to Self-Correct Reasoning
Viaarxiv icon

Reinforcement Learning from Reflective Feedback (RLRF): Aligning and Improving LLMs via Fine-Grained Self-Reflection

Add code
Mar 21, 2024
Figure 1 for Reinforcement Learning from Reflective Feedback (RLRF): Aligning and Improving LLMs via Fine-Grained Self-Reflection
Figure 2 for Reinforcement Learning from Reflective Feedback (RLRF): Aligning and Improving LLMs via Fine-Grained Self-Reflection
Figure 3 for Reinforcement Learning from Reflective Feedback (RLRF): Aligning and Improving LLMs via Fine-Grained Self-Reflection
Figure 4 for Reinforcement Learning from Reflective Feedback (RLRF): Aligning and Improving LLMs via Fine-Grained Self-Reflection
Viaarxiv icon

YTCommentQA: Video Question Answerability in Instructional Videos

Add code
Jan 30, 2024
Viaarxiv icon

Projection Regret: Reducing Background Bias for Novelty Detection via Diffusion Models

Add code
Dec 05, 2023
Figure 1 for Projection Regret: Reducing Background Bias for Novelty Detection via Diffusion Models
Figure 2 for Projection Regret: Reducing Background Bias for Novelty Detection via Diffusion Models
Figure 3 for Projection Regret: Reducing Background Bias for Novelty Detection via Diffusion Models
Figure 4 for Projection Regret: Reducing Background Bias for Novelty Detection via Diffusion Models
Viaarxiv icon

Code Models are Zero-shot Precondition Reasoners

Add code
Nov 16, 2023
Figure 1 for Code Models are Zero-shot Precondition Reasoners
Figure 2 for Code Models are Zero-shot Precondition Reasoners
Figure 3 for Code Models are Zero-shot Precondition Reasoners
Figure 4 for Code Models are Zero-shot Precondition Reasoners
Viaarxiv icon