Picture for Jonathan Berant

Jonathan Berant

Shammie

DOLOMITES: Domain-Specific Long-Form Methodical Tasks

Add code
May 09, 2024
Viaarxiv icon

In-Context Learning with Long-Context Models: An In-Depth Exploration

Add code
Apr 30, 2024
Viaarxiv icon

Large Language Models for Psycholinguistic Plausibility Pretesting

Add code
Feb 08, 2024
Viaarxiv icon

Transforming and Combining Rewards for Aligning Large Language Models

Add code
Feb 01, 2024
Viaarxiv icon

Theoretical guarantees on the best-of-n alignment policy

Add code
Jan 03, 2024
Figure 1 for Theoretical guarantees on the best-of-n alignment policy
Figure 2 for Theoretical guarantees on the best-of-n alignment policy
Figure 3 for Theoretical guarantees on the best-of-n alignment policy
Viaarxiv icon

Helping or Herding? Reward Model Ensembles Mitigate but do not Eliminate Reward Hacking

Add code
Dec 21, 2023
Figure 1 for Helping or Herding? Reward Model Ensembles Mitigate but do not Eliminate Reward Hacking
Figure 2 for Helping or Herding? Reward Model Ensembles Mitigate but do not Eliminate Reward Hacking
Figure 3 for Helping or Herding? Reward Model Ensembles Mitigate but do not Eliminate Reward Hacking
Figure 4 for Helping or Herding? Reward Model Ensembles Mitigate but do not Eliminate Reward Hacking
Viaarxiv icon

SEMQA: Semi-Extractive Multi-Source Question Answering

Add code
Nov 08, 2023
Viaarxiv icon

Never Train from Scratch: Fair Comparison of Long-Sequence Models Requires Data-Driven Priors

Add code
Oct 04, 2023
Viaarxiv icon

Making Retrieval-Augmented Language Models Robust to Irrelevant Context

Add code
Oct 02, 2023
Viaarxiv icon

Long-range Language Modeling with Self-retrieval

Add code
Jun 23, 2023
Viaarxiv icon