Picture for Shruti Bhosale

Shruti Bhosale

NLLB Team

Effective Long-Context Scaling of Foundation Models

Add code
Sep 27, 2023
Figure 1 for Effective Long-Context Scaling of Foundation Models
Figure 2 for Effective Long-Context Scaling of Foundation Models
Figure 3 for Effective Long-Context Scaling of Foundation Models
Figure 4 for Effective Long-Context Scaling of Foundation Models
Viaarxiv icon

Llama 2: Open Foundation and Fine-Tuned Chat Models

Add code
Jul 19, 2023
Figure 1 for Llama 2: Open Foundation and Fine-Tuned Chat Models
Figure 2 for Llama 2: Open Foundation and Fine-Tuned Chat Models
Figure 3 for Llama 2: Open Foundation and Fine-Tuned Chat Models
Figure 4 for Llama 2: Open Foundation and Fine-Tuned Chat Models
Viaarxiv icon

Revisiting Machine Translation for Cross-lingual Classification

Add code
May 23, 2023
Figure 1 for Revisiting Machine Translation for Cross-lingual Classification
Figure 2 for Revisiting Machine Translation for Cross-lingual Classification
Figure 3 for Revisiting Machine Translation for Cross-lingual Classification
Figure 4 for Revisiting Machine Translation for Cross-lingual Classification
Viaarxiv icon

Towards MoE Deployment: Mitigating Inefficiencies in Mixture-of-Expert (MoE) Inference

Add code
Mar 10, 2023
Figure 1 for Towards MoE Deployment: Mitigating Inefficiencies in Mixture-of-Expert (MoE) Inference
Figure 2 for Towards MoE Deployment: Mitigating Inefficiencies in Mixture-of-Expert (MoE) Inference
Figure 3 for Towards MoE Deployment: Mitigating Inefficiencies in Mixture-of-Expert (MoE) Inference
Figure 4 for Towards MoE Deployment: Mitigating Inefficiencies in Mixture-of-Expert (MoE) Inference
Viaarxiv icon

Fixing MoE Over-Fitting on Low-Resource Languages in Multilingual Machine Translation

Add code
Dec 15, 2022
Figure 1 for Fixing MoE Over-Fitting on Low-Resource Languages in Multilingual Machine Translation
Figure 2 for Fixing MoE Over-Fitting on Low-Resource Languages in Multilingual Machine Translation
Figure 3 for Fixing MoE Over-Fitting on Low-Resource Languages in Multilingual Machine Translation
Figure 4 for Fixing MoE Over-Fitting on Low-Resource Languages in Multilingual Machine Translation
Viaarxiv icon

Causes and Cures for Interference in Multilingual Translation

Add code
Dec 14, 2022
Figure 1 for Causes and Cures for Interference in Multilingual Translation
Figure 2 for Causes and Cures for Interference in Multilingual Translation
Figure 3 for Causes and Cures for Interference in Multilingual Translation
Figure 4 for Causes and Cures for Interference in Multilingual Translation
Viaarxiv icon

No Language Left Behind: Scaling Human-Centered Machine Translation

Add code
Jul 11, 2022
Figure 1 for No Language Left Behind: Scaling Human-Centered Machine Translation
Figure 2 for No Language Left Behind: Scaling Human-Centered Machine Translation
Figure 3 for No Language Left Behind: Scaling Human-Centered Machine Translation
Figure 4 for No Language Left Behind: Scaling Human-Centered Machine Translation
Viaarxiv icon

Multilingual Machine Translation with Hyper-Adapters

Add code
May 22, 2022
Figure 1 for Multilingual Machine Translation with Hyper-Adapters
Figure 2 for Multilingual Machine Translation with Hyper-Adapters
Figure 3 for Multilingual Machine Translation with Hyper-Adapters
Figure 4 for Multilingual Machine Translation with Hyper-Adapters
Viaarxiv icon

Data Selection Curriculum for Neural Machine Translation

Add code
Mar 25, 2022
Figure 1 for Data Selection Curriculum for Neural Machine Translation
Figure 2 for Data Selection Curriculum for Neural Machine Translation
Figure 3 for Data Selection Curriculum for Neural Machine Translation
Figure 4 for Data Selection Curriculum for Neural Machine Translation
Viaarxiv icon

Efficient Large Scale Language Modeling with Mixtures of Experts

Add code
Dec 20, 2021
Figure 1 for Efficient Large Scale Language Modeling with Mixtures of Experts
Figure 2 for Efficient Large Scale Language Modeling with Mixtures of Experts
Figure 3 for Efficient Large Scale Language Modeling with Mixtures of Experts
Figure 4 for Efficient Large Scale Language Modeling with Mixtures of Experts
Viaarxiv icon