Picture for Max Ryabinin

Max Ryabinin

Mind Your Format: Towards Consistent Evaluation of In-Context Learning Improvements

Add code
Jan 22, 2024
Figure 1 for Mind Your Format: Towards Consistent Evaluation of In-Context Learning Improvements
Figure 2 for Mind Your Format: Towards Consistent Evaluation of In-Context Learning Improvements
Figure 3 for Mind Your Format: Towards Consistent Evaluation of In-Context Learning Improvements
Figure 4 for Mind Your Format: Towards Consistent Evaluation of In-Context Learning Improvements
Viaarxiv icon

Distributed Inference and Fine-tuning of Large Language Models Over The Internet

Add code
Dec 13, 2023
Figure 1 for Distributed Inference and Fine-tuning of Large Language Models Over The Internet
Figure 2 for Distributed Inference and Fine-tuning of Large Language Models Over The Internet
Figure 3 for Distributed Inference and Fine-tuning of Large Language Models Over The Internet
Figure 4 for Distributed Inference and Fine-tuning of Large Language Models Over The Internet
Viaarxiv icon

Hypernymy Understanding Evaluation of Text-to-Image Models via WordNet Hierarchy

Add code
Oct 13, 2023
Figure 1 for Hypernymy Understanding Evaluation of Text-to-Image Models via WordNet Hierarchy
Figure 2 for Hypernymy Understanding Evaluation of Text-to-Image Models via WordNet Hierarchy
Figure 3 for Hypernymy Understanding Evaluation of Text-to-Image Models via WordNet Hierarchy
Figure 4 for Hypernymy Understanding Evaluation of Text-to-Image Models via WordNet Hierarchy
Viaarxiv icon

High-throughput Generative Inference of Large Language Models with a Single GPU

Add code
Mar 13, 2023
Figure 1 for High-throughput Generative Inference of Large Language Models with a Single GPU
Figure 2 for High-throughput Generative Inference of Large Language Models with a Single GPU
Figure 3 for High-throughput Generative Inference of Large Language Models with a Single GPU
Figure 4 for High-throughput Generative Inference of Large Language Models with a Single GPU
Viaarxiv icon

Is This Loss Informative? Speeding Up Textual Inversion with Deterministic Objective Evaluation

Add code
Feb 09, 2023
Viaarxiv icon

SWARM Parallelism: Training Large Models Can Be Surprisingly Communication-Efficient

Add code
Jan 27, 2023
Figure 1 for SWARM Parallelism: Training Large Models Can Be Surprisingly Communication-Efficient
Figure 2 for SWARM Parallelism: Training Large Models Can Be Surprisingly Communication-Efficient
Figure 3 for SWARM Parallelism: Training Large Models Can Be Surprisingly Communication-Efficient
Figure 4 for SWARM Parallelism: Training Large Models Can Be Surprisingly Communication-Efficient
Viaarxiv icon

BLOOM: A 176B-Parameter Open-Access Multilingual Language Model

Add code
Nov 09, 2022
Viaarxiv icon

RuCoLA: Russian Corpus of Linguistic Acceptability

Add code
Oct 23, 2022
Figure 1 for RuCoLA: Russian Corpus of Linguistic Acceptability
Figure 2 for RuCoLA: Russian Corpus of Linguistic Acceptability
Figure 3 for RuCoLA: Russian Corpus of Linguistic Acceptability
Figure 4 for RuCoLA: Russian Corpus of Linguistic Acceptability
Viaarxiv icon

Petals: Collaborative Inference and Fine-tuning of Large Models

Add code
Sep 02, 2022
Figure 1 for Petals: Collaborative Inference and Fine-tuning of Large Models
Figure 2 for Petals: Collaborative Inference and Fine-tuning of Large Models
Figure 3 for Petals: Collaborative Inference and Fine-tuning of Large Models
Figure 4 for Petals: Collaborative Inference and Fine-tuning of Large Models
Viaarxiv icon

Training Transformers Together

Add code
Jul 07, 2022
Figure 1 for Training Transformers Together
Figure 2 for Training Transformers Together
Viaarxiv icon