Picture for Max Ryabinin

Max Ryabinin

Mashup Learning: Faster Finetuning by Remixing Past Checkpoints

Add code
Mar 10, 2026
Viaarxiv icon

Untied Ulysses: Memory-Efficient Context Parallelism via Headwise Chunking

Add code
Feb 24, 2026
Viaarxiv icon

AutoJudge: Judge Decoding Without Manual Annotation

Add code
Apr 28, 2025
Figure 1 for AutoJudge: Judge Decoding Without Manual Annotation
Figure 2 for AutoJudge: Judge Decoding Without Manual Annotation
Figure 3 for AutoJudge: Judge Decoding Without Manual Annotation
Figure 4 for AutoJudge: Judge Decoding Without Manual Annotation
Viaarxiv icon

Multilingual Language Model Pretraining using Machine-translated Data

Add code
Feb 18, 2025
Viaarxiv icon

Towards Best Practices for Open Datasets for LLM Training

Add code
Jan 14, 2025
Viaarxiv icon

Label Privacy in Split Learning for Large Models with Parameter-Efficient Training

Add code
Dec 21, 2024
Viaarxiv icon

RedPajama: an Open Dataset for Training Large Language Models

Add code
Nov 19, 2024
Viaarxiv icon

Multilingual Pretraining Using a Large Corpus Machine-Translated from a Single Source Language

Add code
Oct 31, 2024
Figure 1 for Multilingual Pretraining Using a Large Corpus Machine-Translated from a Single Source Language
Figure 2 for Multilingual Pretraining Using a Large Corpus Machine-Translated from a Single Source Language
Figure 3 for Multilingual Pretraining Using a Large Corpus Machine-Translated from a Single Source Language
Figure 4 for Multilingual Pretraining Using a Large Corpus Machine-Translated from a Single Source Language
Viaarxiv icon

SpecExec: Massively Parallel Speculative Decoding for Interactive LLM Inference on Consumer Devices

Add code
Jun 04, 2024
Figure 1 for SpecExec: Massively Parallel Speculative Decoding for Interactive LLM Inference on Consumer Devices
Figure 2 for SpecExec: Massively Parallel Speculative Decoding for Interactive LLM Inference on Consumer Devices
Figure 3 for SpecExec: Massively Parallel Speculative Decoding for Interactive LLM Inference on Consumer Devices
Figure 4 for SpecExec: Massively Parallel Speculative Decoding for Interactive LLM Inference on Consumer Devices
Viaarxiv icon

The Hallucinations Leaderboard -- An Open Effort to Measure Hallucinations in Large Language Models

Add code
Apr 08, 2024
Viaarxiv icon