Picture for Nadezhda Chirkova

Nadezhda Chirkova

HSE University, Russia

Investigating the potential of Sparse Mixtures-of-Experts for multi-domain neural machine translation

Add code
Jul 01, 2024
Figure 1 for Investigating the potential of Sparse Mixtures-of-Experts for multi-domain neural machine translation
Figure 2 for Investigating the potential of Sparse Mixtures-of-Experts for multi-domain neural machine translation
Figure 3 for Investigating the potential of Sparse Mixtures-of-Experts for multi-domain neural machine translation
Figure 4 for Investigating the potential of Sparse Mixtures-of-Experts for multi-domain neural machine translation
Viaarxiv icon

Retrieval-augmented generation in multilingual settings

Add code
Jul 01, 2024
Figure 1 for Retrieval-augmented generation in multilingual settings
Figure 2 for Retrieval-augmented generation in multilingual settings
Figure 3 for Retrieval-augmented generation in multilingual settings
Figure 4 for Retrieval-augmented generation in multilingual settings
Viaarxiv icon

BERGEN: A Benchmarking Library for Retrieval-Augmented Generation

Add code
Jul 01, 2024
Figure 1 for BERGEN: A Benchmarking Library for Retrieval-Augmented Generation
Figure 2 for BERGEN: A Benchmarking Library for Retrieval-Augmented Generation
Figure 3 for BERGEN: A Benchmarking Library for Retrieval-Augmented Generation
Figure 4 for BERGEN: A Benchmarking Library for Retrieval-Augmented Generation
Viaarxiv icon

Zero-shot cross-lingual transfer in instruction tuning of large language model

Add code
Feb 22, 2024
Figure 1 for Zero-shot cross-lingual transfer in instruction tuning of large language model
Figure 2 for Zero-shot cross-lingual transfer in instruction tuning of large language model
Figure 3 for Zero-shot cross-lingual transfer in instruction tuning of large language model
Figure 4 for Zero-shot cross-lingual transfer in instruction tuning of large language model
Viaarxiv icon

Key ingredients for effective zero-shot cross-lingual knowledge transfer in generative tasks

Add code
Feb 19, 2024
Figure 1 for Key ingredients for effective zero-shot cross-lingual knowledge transfer in generative tasks
Figure 2 for Key ingredients for effective zero-shot cross-lingual knowledge transfer in generative tasks
Figure 3 for Key ingredients for effective zero-shot cross-lingual knowledge transfer in generative tasks
Figure 4 for Key ingredients for effective zero-shot cross-lingual knowledge transfer in generative tasks
Viaarxiv icon

Empirical study of pretrained multilingual language models for zero-shot cross-lingual generation

Add code
Oct 15, 2023
Viaarxiv icon

CodeBPE: Investigating Subtokenization Options for Large Language Model Pretraining on Source Code

Add code
Aug 01, 2023
Figure 1 for CodeBPE: Investigating Subtokenization Options for Large Language Model Pretraining on Source Code
Figure 2 for CodeBPE: Investigating Subtokenization Options for Large Language Model Pretraining on Source Code
Figure 3 for CodeBPE: Investigating Subtokenization Options for Large Language Model Pretraining on Source Code
Figure 4 for CodeBPE: Investigating Subtokenization Options for Large Language Model Pretraining on Source Code
Viaarxiv icon

Should you marginalize over possible tokenizations?

Add code
Jun 30, 2023
Figure 1 for Should you marginalize over possible tokenizations?
Figure 2 for Should you marginalize over possible tokenizations?
Figure 3 for Should you marginalize over possible tokenizations?
Figure 4 for Should you marginalize over possible tokenizations?
Viaarxiv icon

Parameter-Efficient Finetuning of Transformers for Source Code

Add code
Dec 12, 2022
Figure 1 for Parameter-Efficient Finetuning of Transformers for Source Code
Figure 2 for Parameter-Efficient Finetuning of Transformers for Source Code
Figure 3 for Parameter-Efficient Finetuning of Transformers for Source Code
Figure 4 for Parameter-Efficient Finetuning of Transformers for Source Code
Viaarxiv icon

Probing Pretrained Models of Source Code

Add code
Feb 16, 2022
Figure 1 for Probing Pretrained Models of Source Code
Figure 2 for Probing Pretrained Models of Source Code
Figure 3 for Probing Pretrained Models of Source Code
Figure 4 for Probing Pretrained Models of Source Code
Viaarxiv icon