Picture for Max Ryabinin

Max Ryabinin

SpecExec: Massively Parallel Speculative Decoding for Interactive LLM Inference on Consumer Devices

Add code
Jun 04, 2024
Viaarxiv icon

The Hallucinations Leaderboard -- An Open Effort to Measure Hallucinations in Large Language Models

Add code
Apr 08, 2024
Viaarxiv icon

Sequoia: Scalable, Robust, and Hardware-aware Speculative Decoding

Add code
Feb 29, 2024
Viaarxiv icon

Mind Your Format: Towards Consistent Evaluation of In-Context Learning Improvements

Add code
Jan 22, 2024
Viaarxiv icon

Distributed Inference and Fine-tuning of Large Language Models Over The Internet

Add code
Dec 13, 2023
Figure 1 for Distributed Inference and Fine-tuning of Large Language Models Over The Internet
Figure 2 for Distributed Inference and Fine-tuning of Large Language Models Over The Internet
Figure 3 for Distributed Inference and Fine-tuning of Large Language Models Over The Internet
Figure 4 for Distributed Inference and Fine-tuning of Large Language Models Over The Internet
Viaarxiv icon

Hypernymy Understanding Evaluation of Text-to-Image Models via WordNet Hierarchy

Add code
Oct 13, 2023
Figure 1 for Hypernymy Understanding Evaluation of Text-to-Image Models via WordNet Hierarchy
Figure 2 for Hypernymy Understanding Evaluation of Text-to-Image Models via WordNet Hierarchy
Figure 3 for Hypernymy Understanding Evaluation of Text-to-Image Models via WordNet Hierarchy
Figure 4 for Hypernymy Understanding Evaluation of Text-to-Image Models via WordNet Hierarchy
Viaarxiv icon

High-throughput Generative Inference of Large Language Models with a Single GPU

Add code
Mar 13, 2023
Figure 1 for High-throughput Generative Inference of Large Language Models with a Single GPU
Figure 2 for High-throughput Generative Inference of Large Language Models with a Single GPU
Figure 3 for High-throughput Generative Inference of Large Language Models with a Single GPU
Figure 4 for High-throughput Generative Inference of Large Language Models with a Single GPU
Viaarxiv icon

Is This Loss Informative? Speeding Up Textual Inversion with Deterministic Objective Evaluation

Add code
Feb 09, 2023
Figure 1 for Is This Loss Informative? Speeding Up Textual Inversion with Deterministic Objective Evaluation
Figure 2 for Is This Loss Informative? Speeding Up Textual Inversion with Deterministic Objective Evaluation
Figure 3 for Is This Loss Informative? Speeding Up Textual Inversion with Deterministic Objective Evaluation
Figure 4 for Is This Loss Informative? Speeding Up Textual Inversion with Deterministic Objective Evaluation
Viaarxiv icon

SWARM Parallelism: Training Large Models Can Be Surprisingly Communication-Efficient

Add code
Jan 27, 2023
Figure 1 for SWARM Parallelism: Training Large Models Can Be Surprisingly Communication-Efficient
Figure 2 for SWARM Parallelism: Training Large Models Can Be Surprisingly Communication-Efficient
Figure 3 for SWARM Parallelism: Training Large Models Can Be Surprisingly Communication-Efficient
Figure 4 for SWARM Parallelism: Training Large Models Can Be Surprisingly Communication-Efficient
Viaarxiv icon

BLOOM: A 176B-Parameter Open-Access Multilingual Language Model

Add code
Nov 09, 2022
Viaarxiv icon