Alert button
Picture for Todor Mihaylov

Todor Mihaylov

Alert button

Llama 2: Open Foundation and Fine-Tuned Chat Models

Add code
Bookmark button
Alert button
Jul 19, 2023
Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, Dan Bikel, Lukas Blecher, Cristian Canton Ferrer, Moya Chen, Guillem Cucurull, David Esiobu, Jude Fernandes, Jeremy Fu, Wenyin Fu, Brian Fuller, Cynthia Gao, Vedanuj Goswami, Naman Goyal, Anthony Hartshorn, Saghar Hosseini, Rui Hou, Hakan Inan, Marcin Kardas, Viktor Kerkez, Madian Khabsa, Isabel Kloumann, Artem Korenev, Punit Singh Koura, Marie-Anne Lachaux, Thibaut Lavril, Jenya Lee, Diana Liskovich, Yinghai Lu, Yuning Mao, Xavier Martinet, Todor Mihaylov, Pushkar Mishra, Igor Molybog, Yixin Nie, Andrew Poulton, Jeremy Reizenstein, Rashi Rungta, Kalyan Saladi, Alan Schelten, Ruan Silva, Eric Michael Smith, Ranjan Subramanian, Xiaoqing Ellen Tan, Binh Tang, Ross Taylor, Adina Williams, Jian Xiang Kuan, Puxin Xu, Zheng Yan, Iliyan Zarov, Yuchen Zhang, Angela Fan, Melanie Kambadur, Sharan Narang, Aurelien Rodriguez, Robert Stojnic, Sergey Edunov, Thomas Scialom

Figure 1 for Llama 2: Open Foundation and Fine-Tuned Chat Models
Figure 2 for Llama 2: Open Foundation and Fine-Tuned Chat Models
Figure 3 for Llama 2: Open Foundation and Fine-Tuned Chat Models
Figure 4 for Llama 2: Open Foundation and Fine-Tuned Chat Models
Viaarxiv icon

Understanding In-Context Learning via Supportive Pretraining Data

Add code
Bookmark button
Alert button
Jun 26, 2023
Xiaochuang Han, Daniel Simig, Todor Mihaylov, Yulia Tsvetkov, Asli Celikyilmaz, Tianlu Wang

Figure 1 for Understanding In-Context Learning via Supportive Pretraining Data
Figure 2 for Understanding In-Context Learning via Supportive Pretraining Data
Figure 3 for Understanding In-Context Learning via Supportive Pretraining Data
Figure 4 for Understanding In-Context Learning via Supportive Pretraining Data
Viaarxiv icon

bgGLUE: A Bulgarian General Language Understanding Evaluation Benchmark

Add code
Bookmark button
Alert button
Jun 07, 2023
Momchil Hardalov, Pepa Atanasova, Todor Mihaylov, Galia Angelova, Kiril Simov, Petya Osenova, Ves Stoyanov, Ivan Koychev, Preslav Nakov, Dragomir Radev

Figure 1 for bgGLUE: A Bulgarian General Language Understanding Evaluation Benchmark
Figure 2 for bgGLUE: A Bulgarian General Language Understanding Evaluation Benchmark
Figure 3 for bgGLUE: A Bulgarian General Language Understanding Evaluation Benchmark
Figure 4 for bgGLUE: A Bulgarian General Language Understanding Evaluation Benchmark
Viaarxiv icon

Filtering, Distillation, and Hard Negatives for Vision-Language Pre-Training

Add code
Bookmark button
Alert button
Jan 05, 2023
Filip Radenovic, Abhimanyu Dubey, Abhishek Kadian, Todor Mihaylov, Simon Vandenhende, Yash Patel, Yi Wen, Vignesh Ramanathan, Dhruv Mahajan

Figure 1 for Filtering, Distillation, and Hard Negatives for Vision-Language Pre-Training
Figure 2 for Filtering, Distillation, and Hard Negatives for Vision-Language Pre-Training
Figure 3 for Filtering, Distillation, and Hard Negatives for Vision-Language Pre-Training
Figure 4 for Filtering, Distillation, and Hard Negatives for Vision-Language Pre-Training
Viaarxiv icon

OPT-IML: Scaling Language Model Instruction Meta Learning through the Lens of Generalization

Add code
Bookmark button
Alert button
Dec 28, 2022
Srinivasan Iyer, Xi Victoria Lin, Ramakanth Pasunuru, Todor Mihaylov, Daniel Simig, Ping Yu, Kurt Shuster, Tianlu Wang, Qing Liu, Punit Singh Koura, Xian Li, Brian O'Horo, Gabriel Pereyra, Jeff Wang, Christopher Dewan, Asli Celikyilmaz, Luke Zettlemoyer, Ves Stoyanov

Figure 1 for OPT-IML: Scaling Language Model Instruction Meta Learning through the Lens of Generalization
Figure 2 for OPT-IML: Scaling Language Model Instruction Meta Learning through the Lens of Generalization
Figure 3 for OPT-IML: Scaling Language Model Instruction Meta Learning through the Lens of Generalization
Figure 4 for OPT-IML: Scaling Language Model Instruction Meta Learning through the Lens of Generalization
Viaarxiv icon

OPT: Open Pre-trained Transformer Language Models

Add code
Bookmark button
Alert button
May 05, 2022
Susan Zhang, Stephen Roller, Naman Goyal, Mikel Artetxe, Moya Chen, Shuohui Chen, Christopher Dewan, Mona Diab, Xian Li, Xi Victoria Lin, Todor Mihaylov, Myle Ott, Sam Shleifer, Kurt Shuster, Daniel Simig, Punit Singh Koura, Anjali Sridhar, Tianlu Wang, Luke Zettlemoyer

Figure 1 for OPT: Open Pre-trained Transformer Language Models
Figure 2 for OPT: Open Pre-trained Transformer Language Models
Figure 3 for OPT: Open Pre-trained Transformer Language Models
Figure 4 for OPT: Open Pre-trained Transformer Language Models
Viaarxiv icon

Improving In-Context Few-Shot Learning via Self-Supervised Training

Add code
Bookmark button
Alert button
May 03, 2022
Mingda Chen, Jingfei Du, Ramakanth Pasunuru, Todor Mihaylov, Srini Iyer, Veselin Stoyanov, Zornitsa Kozareva

Figure 1 for Improving In-Context Few-Shot Learning via Self-Supervised Training
Figure 2 for Improving In-Context Few-Shot Learning via Self-Supervised Training
Figure 3 for Improving In-Context Few-Shot Learning via Self-Supervised Training
Figure 4 for Improving In-Context Few-Shot Learning via Self-Supervised Training
Viaarxiv icon

Efficient Large Scale Language Modeling with Mixtures of Experts

Add code
Bookmark button
Alert button
Dec 20, 2021
Mikel Artetxe, Shruti Bhosale, Naman Goyal, Todor Mihaylov, Myle Ott, Sam Shleifer, Xi Victoria Lin, Jingfei Du, Srinivasan Iyer, Ramakanth Pasunuru, Giri Anantharaman, Xian Li, Shuohui Chen, Halil Akin, Mandeep Baines, Louis Martin, Xing Zhou, Punit Singh Koura, Brian O'Horo, Jeff Wang, Luke Zettlemoyer, Mona Diab, Zornitsa Kozareva, Ves Stoyanov

Figure 1 for Efficient Large Scale Language Modeling with Mixtures of Experts
Figure 2 for Efficient Large Scale Language Modeling with Mixtures of Experts
Figure 3 for Efficient Large Scale Language Modeling with Mixtures of Experts
Figure 4 for Efficient Large Scale Language Modeling with Mixtures of Experts
Viaarxiv icon

Few-shot Learning with Multilingual Language Models

Add code
Bookmark button
Alert button
Dec 20, 2021
Xi Victoria Lin, Todor Mihaylov, Mikel Artetxe, Tianlu Wang, Shuohui Chen, Daniel Simig, Myle Ott, Naman Goyal, Shruti Bhosale, Jingfei Du, Ramakanth Pasunuru, Sam Shleifer, Punit Singh Koura, Vishrav Chaudhary, Brian O'Horo, Jeff Wang, Luke Zettlemoyer, Zornitsa Kozareva, Mona Diab, Veselin Stoyanov, Xian Li

Figure 1 for Few-shot Learning with Multilingual Language Models
Figure 2 for Few-shot Learning with Multilingual Language Models
Figure 3 for Few-shot Learning with Multilingual Language Models
Figure 4 for Few-shot Learning with Multilingual Language Models
Viaarxiv icon

SUper Team at SemEval-2016 Task 3: Building a feature-rich system for community question answering

Add code
Bookmark button
Alert button
Sep 26, 2021
Tsvetomila Mihaylova, Pepa Gencheva, Martin Boyanov, Ivana Yovcheva, Todor Mihaylov, Momchil Hardalov, Yasen Kiprov, Daniel Balchev, Ivan Koychev, Preslav Nakov, Ivelina Nikolova, Galia Angelova

Figure 1 for SUper Team at SemEval-2016 Task 3: Building a feature-rich system for community question answering
Figure 2 for SUper Team at SemEval-2016 Task 3: Building a feature-rich system for community question answering
Figure 3 for SUper Team at SemEval-2016 Task 3: Building a feature-rich system for community question answering
Figure 4 for SUper Team at SemEval-2016 Task 3: Building a feature-rich system for community question answering
Viaarxiv icon