Picture for Baolin Peng

Baolin Peng

EJ

Collaborative decoding of critical tokens for boosting factuality of large language models

Add code
Feb 28, 2024
Viaarxiv icon

Fine-Grained Self-Endorsement Improves Factuality and Reasoning

Add code
Feb 23, 2024
Viaarxiv icon

Self-Alignment for Factuality: Mitigating Hallucinations in LLMs via Self-Evaluation

Add code
Feb 14, 2024
Viaarxiv icon

Sub-Sentence Encoder: Contrastive Learning of Propositional Semantic Representations

Add code
Nov 07, 2023
Viaarxiv icon

Teaching Language Models to Self-Improve through Interactive Demonstrations

Add code
Oct 20, 2023
Figure 1 for Teaching Language Models to Self-Improve through Interactive Demonstrations
Figure 2 for Teaching Language Models to Self-Improve through Interactive Demonstrations
Figure 3 for Teaching Language Models to Self-Improve through Interactive Demonstrations
Figure 4 for Teaching Language Models to Self-Improve through Interactive Demonstrations
Viaarxiv icon

The Trickle-down Impact of Reward consistency on RLHF

Add code
Sep 28, 2023
Figure 1 for The Trickle-down Impact of Reward consistency on RLHF
Figure 2 for The Trickle-down Impact of Reward consistency on RLHF
Figure 3 for The Trickle-down Impact of Reward consistency on RLHF
Figure 4 for The Trickle-down Impact of Reward consistency on RLHF
Viaarxiv icon

Stabilizing RLHF through Advantage Model and Selective Rehearsal

Add code
Sep 18, 2023
Viaarxiv icon

Do you really follow me? Adversarial Instructions for Evaluating the Robustness of Large Language Models

Add code
Aug 17, 2023
Figure 1 for Do you really follow me? Adversarial Instructions for Evaluating the Robustness of Large Language Models
Figure 2 for Do you really follow me? Adversarial Instructions for Evaluating the Robustness of Large Language Models
Figure 3 for Do you really follow me? Adversarial Instructions for Evaluating the Robustness of Large Language Models
Figure 4 for Do you really follow me? Adversarial Instructions for Evaluating the Robustness of Large Language Models
Viaarxiv icon

Self-Checker: Plug-and-Play Modules for Fact-Checking with Large Language Models

Add code
May 24, 2023
Figure 1 for Self-Checker: Plug-and-Play Modules for Fact-Checking with Large Language Models
Figure 2 for Self-Checker: Plug-and-Play Modules for Fact-Checking with Large Language Models
Figure 3 for Self-Checker: Plug-and-Play Modules for Fact-Checking with Large Language Models
Figure 4 for Self-Checker: Plug-and-Play Modules for Fact-Checking with Large Language Models
Viaarxiv icon

SGP-TOD: Building Task Bots Effortlessly via Schema-Guided LLM Prompting

Add code
May 15, 2023
Figure 1 for SGP-TOD: Building Task Bots Effortlessly via Schema-Guided LLM Prompting
Figure 2 for SGP-TOD: Building Task Bots Effortlessly via Schema-Guided LLM Prompting
Figure 3 for SGP-TOD: Building Task Bots Effortlessly via Schema-Guided LLM Prompting
Figure 4 for SGP-TOD: Building Task Bots Effortlessly via Schema-Guided LLM Prompting
Viaarxiv icon