Picture for Eric Mitchell

Eric Mitchell

Online Adaptation of Language Models with a Memory of Amortized Contexts

Add code
Mar 07, 2024
Figure 1 for Online Adaptation of Language Models with a Memory of Amortized Contexts
Figure 2 for Online Adaptation of Language Models with a Memory of Amortized Contexts
Figure 3 for Online Adaptation of Language Models with a Memory of Amortized Contexts
Figure 4 for Online Adaptation of Language Models with a Memory of Amortized Contexts
Viaarxiv icon

A Critical Evaluation of AI Feedback for Aligning Large Language Models

Add code
Feb 19, 2024
Figure 1 for A Critical Evaluation of AI Feedback for Aligning Large Language Models
Figure 2 for A Critical Evaluation of AI Feedback for Aligning Large Language Models
Figure 3 for A Critical Evaluation of AI Feedback for Aligning Large Language Models
Figure 4 for A Critical Evaluation of AI Feedback for Aligning Large Language Models
Viaarxiv icon

RLVF: Learning from Verbal Feedback without Overgeneralization

Add code
Feb 16, 2024
Viaarxiv icon

Fine-tuning Language Models for Factuality

Add code
Nov 14, 2023
Viaarxiv icon

An Emulator for Fine-Tuning Large Language Models using Small Language Models

Add code
Oct 19, 2023
Figure 1 for An Emulator for Fine-Tuning Large Language Models using Small Language Models
Figure 2 for An Emulator for Fine-Tuning Large Language Models using Small Language Models
Figure 3 for An Emulator for Fine-Tuning Large Language Models using Small Language Models
Figure 4 for An Emulator for Fine-Tuning Large Language Models using Small Language Models
Viaarxiv icon

Identifying and Mitigating the Security Risks of Generative AI

Add code
Aug 28, 2023
Figure 1 for Identifying and Mitigating the Security Risks of Generative AI
Viaarxiv icon

Direct Preference Optimization: Your Language Model is Secretly a Reward Model

Add code
May 29, 2023
Figure 1 for Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Figure 2 for Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Figure 3 for Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Figure 4 for Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Viaarxiv icon

Just Ask for Calibration: Strategies for Eliciting Calibrated Confidence Scores from Language Models Fine-Tuned with Human Feedback

Add code
May 24, 2023
Figure 1 for Just Ask for Calibration: Strategies for Eliciting Calibrated Confidence Scores from Language Models Fine-Tuned with Human Feedback
Figure 2 for Just Ask for Calibration: Strategies for Eliciting Calibrated Confidence Scores from Language Models Fine-Tuned with Human Feedback
Figure 3 for Just Ask for Calibration: Strategies for Eliciting Calibrated Confidence Scores from Language Models Fine-Tuned with Human Feedback
Figure 4 for Just Ask for Calibration: Strategies for Eliciting Calibrated Confidence Scores from Language Models Fine-Tuned with Human Feedback
Viaarxiv icon

Meta-Learning Online Adaptation of Language Models

Add code
May 24, 2023
Figure 1 for Meta-Learning Online Adaptation of Language Models
Figure 2 for Meta-Learning Online Adaptation of Language Models
Figure 3 for Meta-Learning Online Adaptation of Language Models
Figure 4 for Meta-Learning Online Adaptation of Language Models
Viaarxiv icon

RECKONING: Reasoning through Dynamic Knowledge Encoding

Add code
May 23, 2023
Figure 1 for RECKONING: Reasoning through Dynamic Knowledge Encoding
Figure 2 for RECKONING: Reasoning through Dynamic Knowledge Encoding
Figure 3 for RECKONING: Reasoning through Dynamic Knowledge Encoding
Figure 4 for RECKONING: Reasoning through Dynamic Knowledge Encoding
Viaarxiv icon