Alert button
Picture for Luke Zettlemoyer

Luke Zettlemoyer

Alert button

Mega: Moving Average Equipped Gated Attention

Add code
Bookmark button
Alert button
Sep 26, 2022
Xuezhe Ma, Chunting Zhou, Xiang Kong, Junxian He, Liangke Gui, Graham Neubig, Jonathan May, Luke Zettlemoyer

Figure 1 for Mega: Moving Average Equipped Gated Attention
Figure 2 for Mega: Moving Average Equipped Gated Attention
Figure 3 for Mega: Moving Average Equipped Gated Attention
Figure 4 for Mega: Moving Average Equipped Gated Attention
Viaarxiv icon

Selective Annotation Makes Language Models Better Few-Shot Learners

Add code
Bookmark button
Alert button
Sep 05, 2022
Hongjin Su, Jungo Kasai, Chen Henry Wu, Weijia Shi, Tianlu Wang, Jiayi Xin, Rui Zhang, Mari Ostendorf, Luke Zettlemoyer, Noah A. Smith, Tao Yu

Figure 1 for Selective Annotation Makes Language Models Better Few-Shot Learners
Figure 2 for Selective Annotation Makes Language Models Better Few-Shot Learners
Figure 3 for Selective Annotation Makes Language Models Better Few-Shot Learners
Figure 4 for Selective Annotation Makes Language Models Better Few-Shot Learners
Viaarxiv icon

LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale

Add code
Bookmark button
Alert button
Aug 15, 2022
Tim Dettmers, Mike Lewis, Younes Belkada, Luke Zettlemoyer

Figure 1 for LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale
Figure 2 for LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale
Figure 3 for LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale
Figure 4 for LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale
Viaarxiv icon

Branch-Train-Merge: Embarrassingly Parallel Training of Expert Language Models

Add code
Bookmark button
Alert button
Aug 05, 2022
Margaret Li, Suchin Gururangan, Tim Dettmers, Mike Lewis, Tim Althoff, Noah A. Smith, Luke Zettlemoyer

Figure 1 for Branch-Train-Merge: Embarrassingly Parallel Training of Expert Language Models
Figure 2 for Branch-Train-Merge: Embarrassingly Parallel Training of Expert Language Models
Figure 3 for Branch-Train-Merge: Embarrassingly Parallel Training of Expert Language Models
Figure 4 for Branch-Train-Merge: Embarrassingly Parallel Training of Expert Language Models
Viaarxiv icon

Questions Are All You Need to Train a Dense Passage Retriever

Add code
Bookmark button
Alert button
Jun 21, 2022
Devendra Singh Sachan, Mike Lewis, Dani Yogatama, Luke Zettlemoyer, Joelle Pineau, Manzil Zaheer

Figure 1 for Questions Are All You Need to Train a Dense Passage Retriever
Figure 2 for Questions Are All You Need to Train a Dense Passage Retriever
Figure 3 for Questions Are All You Need to Train a Dense Passage Retriever
Figure 4 for Questions Are All You Need to Train a Dense Passage Retriever
Viaarxiv icon

LegoNN: Building Modular Encoder-Decoder Models

Add code
Bookmark button
Alert button
Jun 07, 2022
Siddharth Dalmia, Dmytro Okhonko, Mike Lewis, Sergey Edunov, Shinji Watanabe, Florian Metze, Luke Zettlemoyer, Abdelrahman Mohamed

Figure 1 for LegoNN: Building Modular Encoder-Decoder Models
Figure 2 for LegoNN: Building Modular Encoder-Decoder Models
Figure 3 for LegoNN: Building Modular Encoder-Decoder Models
Figure 4 for LegoNN: Building Modular Encoder-Decoder Models
Viaarxiv icon

Nearest Neighbor Zero-Shot Inference

Add code
Bookmark button
Alert button
May 27, 2022
Weijia Shi, Julian Michael, Suchin Gururangan, Luke Zettlemoyer

Figure 1 for Nearest Neighbor Zero-Shot Inference
Figure 2 for Nearest Neighbor Zero-Shot Inference
Figure 3 for Nearest Neighbor Zero-Shot Inference
Figure 4 for Nearest Neighbor Zero-Shot Inference
Viaarxiv icon

Logical Satisfiability of Counterfactuals for Faithful Explanations in NLI

Add code
Bookmark button
Alert button
May 25, 2022
Suzanna Sia, Anton Belyy, Amjad Almahairi, Madian Khabsa, Luke Zettlemoyer, Lambert Mathias

Figure 1 for Logical Satisfiability of Counterfactuals for Faithful Explanations in NLI
Figure 2 for Logical Satisfiability of Counterfactuals for Faithful Explanations in NLI
Figure 3 for Logical Satisfiability of Counterfactuals for Faithful Explanations in NLI
Figure 4 for Logical Satisfiability of Counterfactuals for Faithful Explanations in NLI
Viaarxiv icon

Analyzing the Mono- and Cross-Lingual Pretraining Dynamics of Multilingual Language Models

Add code
Bookmark button
Alert button
May 24, 2022
Terra Blevins, Hila Gonen, Luke Zettlemoyer

Figure 1 for Analyzing the Mono- and Cross-Lingual Pretraining Dynamics of Multilingual Language Models
Figure 2 for Analyzing the Mono- and Cross-Lingual Pretraining Dynamics of Multilingual Language Models
Figure 3 for Analyzing the Mono- and Cross-Lingual Pretraining Dynamics of Multilingual Language Models
Figure 4 for Analyzing the Mono- and Cross-Lingual Pretraining Dynamics of Multilingual Language Models
Viaarxiv icon

On the Role of Bidirectionality in Language Model Pre-Training

Add code
Bookmark button
Alert button
May 24, 2022
Mikel Artetxe, Jingfei Du, Naman Goyal, Luke Zettlemoyer, Ves Stoyanov

Figure 1 for On the Role of Bidirectionality in Language Model Pre-Training
Figure 2 for On the Role of Bidirectionality in Language Model Pre-Training
Figure 3 for On the Role of Bidirectionality in Language Model Pre-Training
Figure 4 for On the Role of Bidirectionality in Language Model Pre-Training
Viaarxiv icon