Alert button
Picture for Richard Yuanzhe Pang

Richard Yuanzhe Pang

Alert button

Self-Rewarding Language Models

Add code
Bookmark button
Alert button
Jan 18, 2024
Weizhe Yuan, Richard Yuanzhe Pang, Kyunghyun Cho, Sainbayar Sukhbaatar, Jing Xu, Jason Weston

Viaarxiv icon

GPQA: A Graduate-Level Google-Proof Q&A Benchmark

Add code
Bookmark button
Alert button
Nov 20, 2023
David Rein, Betty Li Hou, Asa Cooper Stickland, Jackson Petty, Richard Yuanzhe Pang, Julien Dirani, Julian Michael, Samuel R. Bowman

Viaarxiv icon

Leveraging Implicit Feedback from Deployment Data in Dialogue

Add code
Bookmark button
Alert button
Jul 26, 2023
Richard Yuanzhe Pang, Stephen Roller, Kyunghyun Cho, He He, Jason Weston

Figure 1 for Leveraging Implicit Feedback from Deployment Data in Dialogue
Figure 2 for Leveraging Implicit Feedback from Deployment Data in Dialogue
Figure 3 for Leveraging Implicit Feedback from Deployment Data in Dialogue
Figure 4 for Leveraging Implicit Feedback from Deployment Data in Dialogue
Viaarxiv icon

Testing the General Deductive Reasoning Capacity of Large Language Models Using OOD Examples

Add code
Bookmark button
Alert button
May 24, 2023
Abulhair Saparov, Richard Yuanzhe Pang, Vishakh Padmakumar, Nitish Joshi, Seyed Mehran Kazemi, Najoung Kim, He He

Figure 1 for Testing the General Deductive Reasoning Capacity of Large Language Models Using OOD Examples
Figure 2 for Testing the General Deductive Reasoning Capacity of Large Language Models Using OOD Examples
Figure 3 for Testing the General Deductive Reasoning Capacity of Large Language Models Using OOD Examples
Figure 4 for Testing the General Deductive Reasoning Capacity of Large Language Models Using OOD Examples
Viaarxiv icon

Extrapolative Controlled Sequence Generation via Iterative Refinement

Add code
Bookmark button
Alert button
Mar 08, 2023
Vishakh Padmakumar, Richard Yuanzhe Pang, He He, Ankur P. Parikh

Figure 1 for Extrapolative Controlled Sequence Generation via Iterative Refinement
Figure 2 for Extrapolative Controlled Sequence Generation via Iterative Refinement
Figure 3 for Extrapolative Controlled Sequence Generation via Iterative Refinement
Figure 4 for Extrapolative Controlled Sequence Generation via Iterative Refinement
Viaarxiv icon

Reward Gaming in Conditional Text Generation

Add code
Bookmark button
Alert button
Nov 16, 2022
Richard Yuanzhe Pang, Vishakh Padmakumar, Thibault Sellam, Ankur P. Parikh, He He

Figure 1 for Reward Gaming in Conditional Text Generation
Figure 2 for Reward Gaming in Conditional Text Generation
Figure 3 for Reward Gaming in Conditional Text Generation
Figure 4 for Reward Gaming in Conditional Text Generation
Viaarxiv icon

What Do NLP Researchers Believe? Results of the NLP Community Metasurvey

Add code
Bookmark button
Alert button
Aug 26, 2022
Julian Michael, Ari Holtzman, Alicia Parrish, Aaron Mueller, Alex Wang, Angelica Chen, Divyam Madaan, Nikita Nangia, Richard Yuanzhe Pang, Jason Phang, Samuel R. Bowman

Figure 1 for What Do NLP Researchers Believe? Results of the NLP Community Metasurvey
Figure 2 for What Do NLP Researchers Believe? Results of the NLP Community Metasurvey
Figure 3 for What Do NLP Researchers Believe? Results of the NLP Community Metasurvey
Figure 4 for What Do NLP Researchers Believe? Results of the NLP Community Metasurvey
Viaarxiv icon

SQuALITY: Building a Long-Document Summarization Dataset the Hard Way

Add code
Bookmark button
Alert button
May 23, 2022
Alex Wang, Richard Yuanzhe Pang, Angelica Chen, Jason Phang, Samuel R. Bowman

Figure 1 for SQuALITY: Building a Long-Document Summarization Dataset the Hard Way
Figure 2 for SQuALITY: Building a Long-Document Summarization Dataset the Hard Way
Figure 3 for SQuALITY: Building a Long-Document Summarization Dataset the Hard Way
Figure 4 for SQuALITY: Building a Long-Document Summarization Dataset the Hard Way
Viaarxiv icon

Token Dropping for Efficient BERT Pretraining

Add code
Bookmark button
Alert button
Mar 24, 2022
Le Hou, Richard Yuanzhe Pang, Tianyi Zhou, Yuexin Wu, Xinying Song, Xiaodan Song, Denny Zhou

Figure 1 for Token Dropping for Efficient BERT Pretraining
Figure 2 for Token Dropping for Efficient BERT Pretraining
Figure 3 for Token Dropping for Efficient BERT Pretraining
Figure 4 for Token Dropping for Efficient BERT Pretraining
Viaarxiv icon

Amortized Noisy Channel Neural Machine Translation

Add code
Bookmark button
Alert button
Dec 16, 2021
Richard Yuanzhe Pang, He He, Kyunghyun Cho

Figure 1 for Amortized Noisy Channel Neural Machine Translation
Figure 2 for Amortized Noisy Channel Neural Machine Translation
Figure 3 for Amortized Noisy Channel Neural Machine Translation
Figure 4 for Amortized Noisy Channel Neural Machine Translation
Viaarxiv icon