Picture for Thomas Wang

Thomas Wang

Mixtral of Experts

Add code
Jan 08, 2024
Viaarxiv icon

FinGPT: Large Generative Models for a Small Language

Add code
Nov 03, 2023
Figure 1 for FinGPT: Large Generative Models for a Small Language
Figure 2 for FinGPT: Large Generative Models for a Small Language
Figure 3 for FinGPT: Large Generative Models for a Small Language
Figure 4 for FinGPT: Large Generative Models for a Small Language
Viaarxiv icon

Mistral 7B

Add code
Oct 10, 2023
Viaarxiv icon

OBELISC: An Open Web-Scale Filtered Dataset of Interleaved Image-Text Documents

Add code
Jun 21, 2023
Figure 1 for OBELISC: An Open Web-Scale Filtered Dataset of Interleaved Image-Text Documents
Figure 2 for OBELISC: An Open Web-Scale Filtered Dataset of Interleaved Image-Text Documents
Figure 3 for OBELISC: An Open Web-Scale Filtered Dataset of Interleaved Image-Text Documents
Figure 4 for OBELISC: An Open Web-Scale Filtered Dataset of Interleaved Image-Text Documents
Viaarxiv icon

StarCoder: may the source be with you!

Add code
May 09, 2023
Figure 1 for StarCoder: may the source be with you!
Figure 2 for StarCoder: may the source be with you!
Figure 3 for StarCoder: may the source be with you!
Figure 4 for StarCoder: may the source be with you!
Viaarxiv icon

The BigScience ROOTS Corpus: A 1.6TB Composite Multilingual Dataset

Add code
Mar 07, 2023
Figure 1 for The BigScience ROOTS Corpus: A 1.6TB Composite Multilingual Dataset
Figure 2 for The BigScience ROOTS Corpus: A 1.6TB Composite Multilingual Dataset
Figure 3 for The BigScience ROOTS Corpus: A 1.6TB Composite Multilingual Dataset
Figure 4 for The BigScience ROOTS Corpus: A 1.6TB Composite Multilingual Dataset
Viaarxiv icon

BLOOM: A 176B-Parameter Open-Access Multilingual Language Model

Add code
Nov 09, 2022
Viaarxiv icon

What Language Model to Train if You Have One Million GPU Hours?

Add code
Nov 08, 2022
Figure 1 for What Language Model to Train if You Have One Million GPU Hours?
Figure 2 for What Language Model to Train if You Have One Million GPU Hours?
Figure 3 for What Language Model to Train if You Have One Million GPU Hours?
Figure 4 for What Language Model to Train if You Have One Million GPU Hours?
Viaarxiv icon

Crosslingual Generalization through Multitask Finetuning

Add code
Nov 03, 2022
Viaarxiv icon

What Language Model Architecture and Pretraining Objective Work Best for Zero-Shot Generalization?

Add code
Apr 12, 2022
Figure 1 for What Language Model Architecture and Pretraining Objective Work Best for Zero-Shot Generalization?
Figure 2 for What Language Model Architecture and Pretraining Objective Work Best for Zero-Shot Generalization?
Figure 3 for What Language Model Architecture and Pretraining Objective Work Best for Zero-Shot Generalization?
Figure 4 for What Language Model Architecture and Pretraining Objective Work Best for Zero-Shot Generalization?
Viaarxiv icon