Alert button
Picture for Eric Mitchell

Eric Mitchell

Alert button

Online Adaptation of Language Models with a Memory of Amortized Contexts

Add code
Bookmark button
Alert button
Mar 07, 2024
Jihoon Tack, Jaehyung Kim, Eric Mitchell, Jinwoo Shin, Yee Whye Teh, Jonathan Richard Schwarz

Figure 1 for Online Adaptation of Language Models with a Memory of Amortized Contexts
Figure 2 for Online Adaptation of Language Models with a Memory of Amortized Contexts
Figure 3 for Online Adaptation of Language Models with a Memory of Amortized Contexts
Figure 4 for Online Adaptation of Language Models with a Memory of Amortized Contexts
Viaarxiv icon

A Critical Evaluation of AI Feedback for Aligning Large Language Models

Add code
Bookmark button
Alert button
Feb 19, 2024
Archit Sharma, Sedrick Keh, Eric Mitchell, Chelsea Finn, Kushal Arora, Thomas Kollar

Viaarxiv icon

RLVF: Learning from Verbal Feedback without Overgeneralization

Add code
Bookmark button
Alert button
Feb 16, 2024
Moritz Stephan, Alexander Khazatsky, Eric Mitchell, Annie S Chen, Sheryl Hsu, Archit Sharma, Chelsea Finn

Viaarxiv icon

Fine-tuning Language Models for Factuality

Add code
Bookmark button
Alert button
Nov 14, 2023
Katherine Tian, Eric Mitchell, Huaxiu Yao, Christopher D. Manning, Chelsea Finn

Viaarxiv icon

An Emulator for Fine-Tuning Large Language Models using Small Language Models

Add code
Bookmark button
Alert button
Oct 19, 2023
Eric Mitchell, Rafael Rafailov, Archit Sharma, Chelsea Finn, Christopher D. Manning

Figure 1 for An Emulator for Fine-Tuning Large Language Models using Small Language Models
Figure 2 for An Emulator for Fine-Tuning Large Language Models using Small Language Models
Figure 3 for An Emulator for Fine-Tuning Large Language Models using Small Language Models
Figure 4 for An Emulator for Fine-Tuning Large Language Models using Small Language Models
Viaarxiv icon

Identifying and Mitigating the Security Risks of Generative AI

Add code
Bookmark button
Alert button
Aug 28, 2023
Clark Barrett, Brad Boyd, Ellie Burzstein, Nicholas Carlini, Brad Chen, Jihye Choi, Amrita Roy Chowdhury, Mihai Christodorescu, Anupam Datta, Soheil Feizi, Kathleen Fisher, Tatsunori Hashimoto, Dan Hendrycks, Somesh Jha, Daniel Kang, Florian Kerschbaum, Eric Mitchell, John Mitchell, Zulfikar Ramzan, Khawaja Shams, Dawn Song, Ankur Taly, Diyi Yang

Figure 1 for Identifying and Mitigating the Security Risks of Generative AI
Viaarxiv icon

Direct Preference Optimization: Your Language Model is Secretly a Reward Model

Add code
Bookmark button
Alert button
May 29, 2023
Rafael Rafailov, Archit Sharma, Eric Mitchell, Stefano Ermon, Christopher D. Manning, Chelsea Finn

Figure 1 for Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Figure 2 for Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Figure 3 for Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Figure 4 for Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Viaarxiv icon

Meta-Learning Online Adaptation of Language Models

Add code
Bookmark button
Alert button
May 24, 2023
Nathan Hu, Eric Mitchell, Christopher D. Manning, Chelsea Finn

Figure 1 for Meta-Learning Online Adaptation of Language Models
Figure 2 for Meta-Learning Online Adaptation of Language Models
Figure 3 for Meta-Learning Online Adaptation of Language Models
Figure 4 for Meta-Learning Online Adaptation of Language Models
Viaarxiv icon

Just Ask for Calibration: Strategies for Eliciting Calibrated Confidence Scores from Language Models Fine-Tuned with Human Feedback

Add code
Bookmark button
Alert button
May 24, 2023
Katherine Tian, Eric Mitchell, Allan Zhou, Archit Sharma, Rafael Rafailov, Huaxiu Yao, Chelsea Finn, Christopher D. Manning

Figure 1 for Just Ask for Calibration: Strategies for Eliciting Calibrated Confidence Scores from Language Models Fine-Tuned with Human Feedback
Figure 2 for Just Ask for Calibration: Strategies for Eliciting Calibrated Confidence Scores from Language Models Fine-Tuned with Human Feedback
Figure 3 for Just Ask for Calibration: Strategies for Eliciting Calibrated Confidence Scores from Language Models Fine-Tuned with Human Feedback
Figure 4 for Just Ask for Calibration: Strategies for Eliciting Calibrated Confidence Scores from Language Models Fine-Tuned with Human Feedback
Viaarxiv icon

RECKONING: Reasoning through Dynamic Knowledge Encoding

Add code
Bookmark button
Alert button
May 23, 2023
Zeming Chen, Gail Weiss, Eric Mitchell, Asli Celikyilmaz, Antoine Bosselut

Figure 1 for RECKONING: Reasoning through Dynamic Knowledge Encoding
Figure 2 for RECKONING: Reasoning through Dynamic Knowledge Encoding
Figure 3 for RECKONING: Reasoning through Dynamic Knowledge Encoding
Figure 4 for RECKONING: Reasoning through Dynamic Knowledge Encoding
Viaarxiv icon