Alert button
Picture for Ahmet Üstün

Ahmet Üstün

Alert button

Back to Basics: Revisiting REINFORCE Style Optimization for Learning from Human Feedback in LLMs

Add code
Bookmark button
Alert button
Feb 26, 2024
Arash Ahmadian, Chris Cremer, Matthias Gallé, Marzieh Fadaee, Julia Kreutzer, Olivier Pietquin, Ahmet Üstün, Sara Hooker

Viaarxiv icon

Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model

Add code
Bookmark button
Alert button
Feb 12, 2024
Ahmet Üstün, Viraat Aryabumi, Zheng-Xin Yong, Wei-Yin Ko, Daniel D'souza, Gbemileke Onilude, Neel Bhandari, Shivalika Singh, Hui-Lee Ooi, Amr Kayid, Freddie Vargus, Phil Blunsom, Shayne Longpre, Niklas Muennighoff, Marzieh Fadaee, Julia Kreutzer, Sara Hooker

Viaarxiv icon

Aya Dataset: An Open-Access Collection for Multilingual Instruction Tuning

Add code
Bookmark button
Alert button
Feb 09, 2024
Shivalika Singh, Freddie Vargus, Daniel Dsouza, Börje F. Karlsson, Abinaya Mahendiran, Wei-Yin Ko, Herumb Shandilya, Jay Patel, Deividas Mataciunas, Laura OMahony, Mike Zhang, Ramith Hettiarachchi, Joseph Wilson, Marina Machado, Luisa Souza Moura, Dominik Krzemiński, Hakimeh Fadaei, Irem Ergün, Ifeoma Okoh, Aisha Alaagib, Oshan Mudannayake, Zaid Alyafeai, Vu Minh Chien, Sebastian Ruder, Surya Guthikonda, Emad A. Alghamdi, Sebastian Gehrmann, Niklas Muennighoff, Max Bartolo, Julia Kreutzer, Ahmet Üstün, Marzieh Fadaee, Sara Hooker

Viaarxiv icon

Pushing Mixture of Experts to the Limit: Extremely Parameter Efficient MoE for Instruction Tuning

Add code
Bookmark button
Alert button
Sep 11, 2023
Ted Zadouri, Ahmet Üstün, Arash Ahmadian, Beyza Ermiş, Acyr Locatelli, Sara Hooker

Figure 1 for Pushing Mixture of Experts to the Limit: Extremely Parameter Efficient MoE for Instruction Tuning
Figure 2 for Pushing Mixture of Experts to the Limit: Extremely Parameter Efficient MoE for Instruction Tuning
Figure 3 for Pushing Mixture of Experts to the Limit: Extremely Parameter Efficient MoE for Instruction Tuning
Figure 4 for Pushing Mixture of Experts to the Limit: Extremely Parameter Efficient MoE for Instruction Tuning
Viaarxiv icon

When Less is More: Investigating Data Pruning for Pretraining LLMs at Scale

Add code
Bookmark button
Alert button
Sep 08, 2023
Max Marion, Ahmet Üstün, Luiza Pozzobon, Alex Wang, Marzieh Fadaee, Sara Hooker

Viaarxiv icon

Intriguing Properties of Quantization at Scale

Add code
Bookmark button
Alert button
May 30, 2023
Arash Ahmadian, Saurabh Dash, Hongyu Chen, Bharat Venkitesh, Stephen Gou, Phil Blunsom, Ahmet Üstün, Sara Hooker

Figure 1 for Intriguing Properties of Quantization at Scale
Figure 2 for Intriguing Properties of Quantization at Scale
Figure 3 for Intriguing Properties of Quantization at Scale
Figure 4 for Intriguing Properties of Quantization at Scale
Viaarxiv icon

Hyper-X: A Unified Hypernetwork for Multi-Task Multilingual Transfer

Add code
Bookmark button
Alert button
May 24, 2022
Ahmet Üstün, Arianna Bisazza, Gosse Bouma, Gertjan van Noord, Sebastian Ruder

Figure 1 for Hyper-X: A Unified Hypernetwork for Multi-Task Multilingual Transfer
Figure 2 for Hyper-X: A Unified Hypernetwork for Multi-Task Multilingual Transfer
Figure 3 for Hyper-X: A Unified Hypernetwork for Multi-Task Multilingual Transfer
Figure 4 for Hyper-X: A Unified Hypernetwork for Multi-Task Multilingual Transfer
Viaarxiv icon

When does Parameter-Efficient Transfer Learning Work for Machine Translation?

Add code
Bookmark button
Alert button
May 23, 2022
Ahmet Üstün, Asa Cooper Stickland

Figure 1 for When does Parameter-Efficient Transfer Learning Work for Machine Translation?
Figure 2 for When does Parameter-Efficient Transfer Learning Work for Machine Translation?
Figure 3 for When does Parameter-Efficient Transfer Learning Work for Machine Translation?
Figure 4 for When does Parameter-Efficient Transfer Learning Work for Machine Translation?
Viaarxiv icon

Multilingual Unsupervised Neural Machine Translation with Denoising Adapters

Add code
Bookmark button
Alert button
Oct 20, 2021
Ahmet Üstün, Alexandre Bérard, Laurent Besacier, Matthias Gallé

Figure 1 for Multilingual Unsupervised Neural Machine Translation with Denoising Adapters
Figure 2 for Multilingual Unsupervised Neural Machine Translation with Denoising Adapters
Figure 3 for Multilingual Unsupervised Neural Machine Translation with Denoising Adapters
Figure 4 for Multilingual Unsupervised Neural Machine Translation with Denoising Adapters
Viaarxiv icon

Unsupervised Translation of German--Lower Sorbian: Exploring Training and Novel Transfer Methods on a Low-Resource Language

Add code
Bookmark button
Alert button
Sep 24, 2021
Lukas Edman, Ahmet Üstün, Antonio Toral, Gertjan van Noord

Figure 1 for Unsupervised Translation of German--Lower Sorbian: Exploring Training and Novel Transfer Methods on a Low-Resource Language
Figure 2 for Unsupervised Translation of German--Lower Sorbian: Exploring Training and Novel Transfer Methods on a Low-Resource Language
Figure 3 for Unsupervised Translation of German--Lower Sorbian: Exploring Training and Novel Transfer Methods on a Low-Resource Language
Figure 4 for Unsupervised Translation of German--Lower Sorbian: Exploring Training and Novel Transfer Methods on a Low-Resource Language
Viaarxiv icon