Alert button
Picture for Hugo Touvron

Hugo Touvron

Alert button

Code Llama: Open Foundation Models for Code

Aug 25, 2023
Baptiste Rozière, Jonas Gehring, Fabian Gloeckle, Sten Sootla, Itai Gat, Xiaoqing Ellen Tan, Yossi Adi, Jingyu Liu, Tal Remez, Jérémy Rapin, Artyom Kozhevnikov, Ivan Evtimov, Joanna Bitton, Manish Bhatt, Cristian Canton Ferrer, Aaron Grattafiori, Wenhan Xiong, Alexandre Défossez, Jade Copet, Faisal Azhar, Hugo Touvron, Louis Martin, Nicolas Usunier, Thomas Scialom, Gabriel Synnaeve

Figure 1 for Code Llama: Open Foundation Models for Code
Figure 2 for Code Llama: Open Foundation Models for Code
Figure 3 for Code Llama: Open Foundation Models for Code
Figure 4 for Code Llama: Open Foundation Models for Code
Viaarxiv icon

Llama 2: Open Foundation and Fine-Tuned Chat Models

Jul 19, 2023
Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, Dan Bikel, Lukas Blecher, Cristian Canton Ferrer, Moya Chen, Guillem Cucurull, David Esiobu, Jude Fernandes, Jeremy Fu, Wenyin Fu, Brian Fuller, Cynthia Gao, Vedanuj Goswami, Naman Goyal, Anthony Hartshorn, Saghar Hosseini, Rui Hou, Hakan Inan, Marcin Kardas, Viktor Kerkez, Madian Khabsa, Isabel Kloumann, Artem Korenev, Punit Singh Koura, Marie-Anne Lachaux, Thibaut Lavril, Jenya Lee, Diana Liskovich, Yinghai Lu, Yuning Mao, Xavier Martinet, Todor Mihaylov, Pushkar Mishra, Igor Molybog, Yixin Nie, Andrew Poulton, Jeremy Reizenstein, Rashi Rungta, Kalyan Saladi, Alan Schelten, Ruan Silva, Eric Michael Smith, Ranjan Subramanian, Xiaoqing Ellen Tan, Binh Tang, Ross Taylor, Adina Williams, Jian Xiang Kuan, Puxin Xu, Zheng Yan, Iliyan Zarov, Yuchen Zhang, Angela Fan, Melanie Kambadur, Sharan Narang, Aurelien Rodriguez, Robert Stojnic, Sergey Edunov, Thomas Scialom

Figure 1 for Llama 2: Open Foundation and Fine-Tuned Chat Models
Figure 2 for Llama 2: Open Foundation and Fine-Tuned Chat Models
Figure 3 for Llama 2: Open Foundation and Fine-Tuned Chat Models
Figure 4 for Llama 2: Open Foundation and Fine-Tuned Chat Models
Viaarxiv icon

LLaMA: Open and Efficient Foundation Language Models

Feb 27, 2023
Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timothée Lacroix, Baptiste Rozière, Naman Goyal, Eric Hambro, Faisal Azhar, Aurelien Rodriguez, Armand Joulin, Edouard Grave, Guillaume Lample

Figure 1 for LLaMA: Open and Efficient Foundation Language Models
Figure 2 for LLaMA: Open and Efficient Foundation Language Models
Figure 3 for LLaMA: Open and Efficient Foundation Language Models
Figure 4 for LLaMA: Open and Efficient Foundation Language Models
Viaarxiv icon

Co-training $2^L$ Submodels for Visual Recognition

Dec 09, 2022
Hugo Touvron, Matthieu Cord, Maxime Oquab, Piotr Bojanowski, Jakob Verbeek, Hervé Jégou

Figure 1 for Co-training $2^L$ Submodels for Visual Recognition
Figure 2 for Co-training $2^L$ Submodels for Visual Recognition
Figure 3 for Co-training $2^L$ Submodels for Visual Recognition
Figure 4 for Co-training $2^L$ Submodels for Visual Recognition
Viaarxiv icon

DeiT III: Revenge of the ViT

Apr 14, 2022
Hugo Touvron, Matthieu Cord, Hervé Jégou

Figure 1 for DeiT III: Revenge of the ViT
Figure 2 for DeiT III: Revenge of the ViT
Figure 3 for DeiT III: Revenge of the ViT
Figure 4 for DeiT III: Revenge of the ViT
Viaarxiv icon

Three things everyone should know about Vision Transformers

Mar 18, 2022
Hugo Touvron, Matthieu Cord, Alaaeldin El-Nouby, Jakob Verbeek, Hervé Jégou

Figure 1 for Three things everyone should know about Vision Transformers
Figure 2 for Three things everyone should know about Vision Transformers
Figure 3 for Three things everyone should know about Vision Transformers
Figure 4 for Three things everyone should know about Vision Transformers
Viaarxiv icon

Augmenting Convolutional networks with attention-based aggregation

Dec 27, 2021
Hugo Touvron, Matthieu Cord, Alaaeldin El-Nouby, Piotr Bojanowski, Armand Joulin, Gabriel Synnaeve, Hervé Jégou

Figure 1 for Augmenting Convolutional networks with attention-based aggregation
Figure 2 for Augmenting Convolutional networks with attention-based aggregation
Figure 3 for Augmenting Convolutional networks with attention-based aggregation
Figure 4 for Augmenting Convolutional networks with attention-based aggregation
Viaarxiv icon

Are Large-scale Datasets Necessary for Self-Supervised Pre-training?

Dec 20, 2021
Alaaeldin El-Nouby, Gautier Izacard, Hugo Touvron, Ivan Laptev, Hervé Jegou, Edouard Grave

Figure 1 for Are Large-scale Datasets Necessary for Self-Supervised Pre-training?
Figure 2 for Are Large-scale Datasets Necessary for Self-Supervised Pre-training?
Figure 3 for Are Large-scale Datasets Necessary for Self-Supervised Pre-training?
Figure 4 for Are Large-scale Datasets Necessary for Self-Supervised Pre-training?
Viaarxiv icon

ResNet strikes back: An improved training procedure in timm

Oct 01, 2021
Ross Wightman, Hugo Touvron, Hervé Jégou

Figure 1 for ResNet strikes back: An improved training procedure in timm
Figure 2 for ResNet strikes back: An improved training procedure in timm
Figure 3 for ResNet strikes back: An improved training procedure in timm
Figure 4 for ResNet strikes back: An improved training procedure in timm
Viaarxiv icon