Alert button
Picture for Mike Lewis

Mike Lewis

Alert button

In-Context Pretraining: Language Modeling Beyond Document Boundaries

Add code
Bookmark button
Alert button
Oct 20, 2023
Weijia Shi, Sewon Min, Maria Lomeli, Chunting Zhou, Margaret Li, Xi Victoria Lin, Noah A. Smith, Luke Zettlemoyer, Scott Yih, Mike Lewis

Figure 1 for In-Context Pretraining: Language Modeling Beyond Document Boundaries
Figure 2 for In-Context Pretraining: Language Modeling Beyond Document Boundaries
Figure 3 for In-Context Pretraining: Language Modeling Beyond Document Boundaries
Figure 4 for In-Context Pretraining: Language Modeling Beyond Document Boundaries
Viaarxiv icon

RA-DIT: Retrieval-Augmented Dual Instruction Tuning

Add code
Bookmark button
Alert button
Oct 08, 2023
Xi Victoria Lin, Xilun Chen, Mingda Chen, Weijia Shi, Maria Lomeli, Rich James, Pedro Rodriguez, Jacob Kahn, Gergely Szilvasy, Mike Lewis, Luke Zettlemoyer, Scott Yih

Figure 1 for RA-DIT: Retrieval-Augmented Dual Instruction Tuning
Figure 2 for RA-DIT: Retrieval-Augmented Dual Instruction Tuning
Figure 3 for RA-DIT: Retrieval-Augmented Dual Instruction Tuning
Figure 4 for RA-DIT: Retrieval-Augmented Dual Instruction Tuning
Viaarxiv icon

Efficient Streaming Language Models with Attention Sinks

Add code
Bookmark button
Alert button
Sep 29, 2023
Guangxuan Xiao, Yuandong Tian, Beidi Chen, Song Han, Mike Lewis

Figure 1 for Efficient Streaming Language Models with Attention Sinks
Figure 2 for Efficient Streaming Language Models with Attention Sinks
Figure 3 for Efficient Streaming Language Models with Attention Sinks
Figure 4 for Efficient Streaming Language Models with Attention Sinks
Viaarxiv icon

Contrastive Decoding Improves Reasoning in Large Language Models

Add code
Bookmark button
Alert button
Sep 29, 2023
Sean O'Brien, Mike Lewis

Viaarxiv icon

Effective Long-Context Scaling of Foundation Models

Add code
Bookmark button
Alert button
Sep 27, 2023
Wenhan Xiong, Jingyu Liu, Igor Molybog, Hejia Zhang, Prajjwal Bhargava, Rui Hou, Louis Martin, Rashi Rungta, Karthik Abinav Sankararaman, Barlas Oguz, Madian Khabsa, Han Fang, Yashar Mehdad, Sharan Narang, Kshitiz Malik, Angela Fan, Shruti Bhosale, Sergey Edunov, Mike Lewis, Sinong Wang, Hao Ma

Figure 1 for Effective Long-Context Scaling of Foundation Models
Figure 2 for Effective Long-Context Scaling of Foundation Models
Figure 3 for Effective Long-Context Scaling of Foundation Models
Figure 4 for Effective Long-Context Scaling of Foundation Models
Viaarxiv icon

Self-Alignment with Instruction Backtranslation

Add code
Bookmark button
Alert button
Aug 14, 2023
Xian Li, Ping Yu, Chunting Zhou, Timo Schick, Luke Zettlemoyer, Omer Levy, Jason Weston, Mike Lewis

Figure 1 for Self-Alignment with Instruction Backtranslation
Figure 2 for Self-Alignment with Instruction Backtranslation
Figure 3 for Self-Alignment with Instruction Backtranslation
Figure 4 for Self-Alignment with Instruction Backtranslation
Viaarxiv icon

Trusting Your Evidence: Hallucinate Less with Context-aware Decoding

Add code
Bookmark button
Alert button
May 24, 2023
Weijia Shi, Xiaochuang Han, Mike Lewis, Yulia Tsvetkov, Luke Zettlemoyer, Scott Wen-tau Yih

Figure 1 for Trusting Your Evidence: Hallucinate Less with Context-aware Decoding
Figure 2 for Trusting Your Evidence: Hallucinate Less with Context-aware Decoding
Figure 3 for Trusting Your Evidence: Hallucinate Less with Context-aware Decoding
Figure 4 for Trusting Your Evidence: Hallucinate Less with Context-aware Decoding
Viaarxiv icon