Alert button
Picture for Daniel Simig

Daniel Simig

Alert button

D4: Improving LLM Pretraining via Document De-Duplication and Diversification

Add code
Bookmark button
Alert button
Aug 23, 2023
Kushal Tirumala, Daniel Simig, Armen Aghajanyan, Ari S. Morcos

Figure 1 for D4: Improving LLM Pretraining via Document De-Duplication and Diversification
Figure 2 for D4: Improving LLM Pretraining via Document De-Duplication and Diversification
Figure 3 for D4: Improving LLM Pretraining via Document De-Duplication and Diversification
Figure 4 for D4: Improving LLM Pretraining via Document De-Duplication and Diversification
Viaarxiv icon

Understanding In-Context Learning via Supportive Pretraining Data

Add code
Bookmark button
Alert button
Jun 26, 2023
Xiaochuang Han, Daniel Simig, Todor Mihaylov, Yulia Tsvetkov, Asli Celikyilmaz, Tianlu Wang

Figure 1 for Understanding In-Context Learning via Supportive Pretraining Data
Figure 2 for Understanding In-Context Learning via Supportive Pretraining Data
Figure 3 for Understanding In-Context Learning via Supportive Pretraining Data
Figure 4 for Understanding In-Context Learning via Supportive Pretraining Data
Viaarxiv icon

Evaluating end-to-end entity linking on domain-specific knowledge bases: Learning about ancient technologies from museum collections

Add code
Bookmark button
Alert button
May 23, 2023
Sebastian Cadavid-Sanchez, Khalil Kacem, Rafael Aparecido Martins Frade, Johannes Boehm, Thomas Chaney, Danial Lashkari, Daniel Simig

Figure 1 for Evaluating end-to-end entity linking on domain-specific knowledge bases: Learning about ancient technologies from museum collections
Figure 2 for Evaluating end-to-end entity linking on domain-specific knowledge bases: Learning about ancient technologies from museum collections
Figure 3 for Evaluating end-to-end entity linking on domain-specific knowledge bases: Learning about ancient technologies from museum collections
Figure 4 for Evaluating end-to-end entity linking on domain-specific knowledge bases: Learning about ancient technologies from museum collections
Viaarxiv icon

OPT-IML: Scaling Language Model Instruction Meta Learning through the Lens of Generalization

Add code
Bookmark button
Alert button
Dec 28, 2022
Srinivasan Iyer, Xi Victoria Lin, Ramakanth Pasunuru, Todor Mihaylov, Daniel Simig, Ping Yu, Kurt Shuster, Tianlu Wang, Qing Liu, Punit Singh Koura, Xian Li, Brian O'Horo, Gabriel Pereyra, Jeff Wang, Christopher Dewan, Asli Celikyilmaz, Luke Zettlemoyer, Ves Stoyanov

Figure 1 for OPT-IML: Scaling Language Model Instruction Meta Learning through the Lens of Generalization
Figure 2 for OPT-IML: Scaling Language Model Instruction Meta Learning through the Lens of Generalization
Figure 3 for OPT-IML: Scaling Language Model Instruction Meta Learning through the Lens of Generalization
Figure 4 for OPT-IML: Scaling Language Model Instruction Meta Learning through the Lens of Generalization
Viaarxiv icon

Text Characterization Toolkit

Add code
Bookmark button
Alert button
Oct 04, 2022
Daniel Simig, Tianlu Wang, Verna Dankers, Peter Henderson, Khuyagbaatar Batsuren, Dieuwke Hupkes, Mona Diab

Figure 1 for Text Characterization Toolkit
Figure 2 for Text Characterization Toolkit
Figure 3 for Text Characterization Toolkit
Figure 4 for Text Characterization Toolkit
Viaarxiv icon

Open Vocabulary Extreme Classification Using Generative Models

Add code
Bookmark button
Alert button
May 12, 2022
Daniel Simig, Fabio Petroni, Pouya Yanki, Kashyap Popat, Christina Du, Sebastian Riedel, Majid Yazdani

Figure 1 for Open Vocabulary Extreme Classification Using Generative Models
Figure 2 for Open Vocabulary Extreme Classification Using Generative Models
Figure 3 for Open Vocabulary Extreme Classification Using Generative Models
Figure 4 for Open Vocabulary Extreme Classification Using Generative Models
Viaarxiv icon

OPT: Open Pre-trained Transformer Language Models

Add code
Bookmark button
Alert button
May 05, 2022
Susan Zhang, Stephen Roller, Naman Goyal, Mikel Artetxe, Moya Chen, Shuohui Chen, Christopher Dewan, Mona Diab, Xian Li, Xi Victoria Lin, Todor Mihaylov, Myle Ott, Sam Shleifer, Kurt Shuster, Daniel Simig, Punit Singh Koura, Anjali Sridhar, Tianlu Wang, Luke Zettlemoyer

Figure 1 for OPT: Open Pre-trained Transformer Language Models
Figure 2 for OPT: Open Pre-trained Transformer Language Models
Figure 3 for OPT: Open Pre-trained Transformer Language Models
Figure 4 for OPT: Open Pre-trained Transformer Language Models
Viaarxiv icon

Few-shot Learning with Multilingual Language Models

Add code
Bookmark button
Alert button
Dec 20, 2021
Xi Victoria Lin, Todor Mihaylov, Mikel Artetxe, Tianlu Wang, Shuohui Chen, Daniel Simig, Myle Ott, Naman Goyal, Shruti Bhosale, Jingfei Du, Ramakanth Pasunuru, Sam Shleifer, Punit Singh Koura, Vishrav Chaudhary, Brian O'Horo, Jeff Wang, Luke Zettlemoyer, Zornitsa Kozareva, Mona Diab, Veselin Stoyanov, Xian Li

Figure 1 for Few-shot Learning with Multilingual Language Models
Figure 2 for Few-shot Learning with Multilingual Language Models
Figure 3 for Few-shot Learning with Multilingual Language Models
Figure 4 for Few-shot Learning with Multilingual Language Models
Viaarxiv icon