Alert button
Picture for Mike Lewis

Mike Lewis

Alert button

8-bit Optimizers via Block-wise Quantization

Add code
Bookmark button
Alert button
Oct 06, 2021
Tim Dettmers, Mike Lewis, Sam Shleifer, Luke Zettlemoyer

Figure 1 for 8-bit Optimizers via Block-wise Quantization
Figure 2 for 8-bit Optimizers via Block-wise Quantization
Figure 3 for 8-bit Optimizers via Block-wise Quantization
Figure 4 for 8-bit Optimizers via Block-wise Quantization
Viaarxiv icon

Train Short, Test Long: Attention with Linear Biases Enables Input Length Extrapolation

Add code
Bookmark button
Alert button
Aug 27, 2021
Ofir Press, Noah A. Smith, Mike Lewis

Figure 1 for Train Short, Test Long: Attention with Linear Biases Enables Input Length Extrapolation
Figure 2 for Train Short, Test Long: Attention with Linear Biases Enables Input Length Extrapolation
Figure 3 for Train Short, Test Long: Attention with Linear Biases Enables Input Length Extrapolation
Figure 4 for Train Short, Test Long: Attention with Linear Biases Enables Input Length Extrapolation
Viaarxiv icon

DEMix Layers: Disentangling Domains for Modular Language Modeling

Add code
Bookmark button
Alert button
Aug 20, 2021
Suchin Gururangan, Mike Lewis, Ari Holtzman, Noah A. Smith, Luke Zettlemoyer

Figure 1 for DEMix Layers: Disentangling Domains for Modular Language Modeling
Figure 2 for DEMix Layers: Disentangling Domains for Modular Language Modeling
Figure 3 for DEMix Layers: Disentangling Domains for Modular Language Modeling
Figure 4 for DEMix Layers: Disentangling Domains for Modular Language Modeling
Viaarxiv icon

Noisy Channel Language Model Prompting for Few-Shot Text Classification

Add code
Bookmark button
Alert button
Aug 15, 2021
Sewon Min, Mike Lewis, Hannaneh Hajishirzi, Luke Zettlemoyer

Figure 1 for Noisy Channel Language Model Prompting for Few-Shot Text Classification
Figure 2 for Noisy Channel Language Model Prompting for Few-Shot Text Classification
Figure 3 for Noisy Channel Language Model Prompting for Few-Shot Text Classification
Figure 4 for Noisy Channel Language Model Prompting for Few-Shot Text Classification
Viaarxiv icon

HTLM: Hyper-Text Pre-Training and Prompting of Language Models

Add code
Bookmark button
Alert button
Jul 14, 2021
Armen Aghajanyan, Dmytro Okhonko, Mike Lewis, Mandar Joshi, Hu Xu, Gargi Ghosh, Luke Zettlemoyer

Figure 1 for HTLM: Hyper-Text Pre-Training and Prompting of Language Models
Figure 2 for HTLM: Hyper-Text Pre-Training and Prompting of Language Models
Figure 3 for HTLM: Hyper-Text Pre-Training and Prompting of Language Models
Figure 4 for HTLM: Hyper-Text Pre-Training and Prompting of Language Models
Viaarxiv icon

Question Answering Infused Pre-training of General-Purpose Contextualized Representations

Add code
Bookmark button
Alert button
Jun 15, 2021
Robin Jia, Mike Lewis, Luke Zettlemoyer

Figure 1 for Question Answering Infused Pre-training of General-Purpose Contextualized Representations
Figure 2 for Question Answering Infused Pre-training of General-Purpose Contextualized Representations
Figure 3 for Question Answering Infused Pre-training of General-Purpose Contextualized Representations
Figure 4 for Question Answering Infused Pre-training of General-Purpose Contextualized Representations
Viaarxiv icon

Multitasking Inhibits Semantic Drift

Add code
Bookmark button
Alert button
Apr 15, 2021
Athul Paul Jacob, Mike Lewis, Jacob Andreas

Figure 1 for Multitasking Inhibits Semantic Drift
Figure 2 for Multitasking Inhibits Semantic Drift
Figure 3 for Multitasking Inhibits Semantic Drift
Figure 4 for Multitasking Inhibits Semantic Drift
Viaarxiv icon

BASE Layers: Simplifying Training of Large, Sparse Models

Add code
Bookmark button
Alert button
Mar 30, 2021
Mike Lewis, Shruti Bhosale, Tim Dettmers, Naman Goyal, Luke Zettlemoyer

Figure 1 for BASE Layers: Simplifying Training of Large, Sparse Models
Figure 2 for BASE Layers: Simplifying Training of Large, Sparse Models
Figure 3 for BASE Layers: Simplifying Training of Large, Sparse Models
Figure 4 for BASE Layers: Simplifying Training of Large, Sparse Models
Viaarxiv icon