Picture for Christopher Ré

Christopher Ré

Department of Computer Science, Stanford University

Restructuring Vector Quantization with the Rotation Trick

Add code
Oct 08, 2024
Figure 1 for Restructuring Vector Quantization with the Rotation Trick
Figure 2 for Restructuring Vector Quantization with the Rotation Trick
Figure 3 for Restructuring Vector Quantization with the Rotation Trick
Figure 4 for Restructuring Vector Quantization with the Rotation Trick
Viaarxiv icon

Cookbook: A framework for improving LLM generative abilities via programmatic data generating templates

Add code
Oct 07, 2024
Figure 1 for Cookbook: A framework for improving LLM generative abilities via programmatic data generating templates
Figure 2 for Cookbook: A framework for improving LLM generative abilities via programmatic data generating templates
Figure 3 for Cookbook: A framework for improving LLM generative abilities via programmatic data generating templates
Figure 4 for Cookbook: A framework for improving LLM generative abilities via programmatic data generating templates
Viaarxiv icon

Large Language Monkeys: Scaling Inference Compute with Repeated Sampling

Add code
Jul 31, 2024
Figure 1 for Large Language Monkeys: Scaling Inference Compute with Repeated Sampling
Figure 2 for Large Language Monkeys: Scaling Inference Compute with Repeated Sampling
Figure 3 for Large Language Monkeys: Scaling Inference Compute with Repeated Sampling
Figure 4 for Large Language Monkeys: Scaling Inference Compute with Repeated Sampling
Viaarxiv icon

Just read twice: closing the recall gap for recurrent language models

Add code
Jul 07, 2024
Figure 1 for Just read twice: closing the recall gap for recurrent language models
Figure 2 for Just read twice: closing the recall gap for recurrent language models
Figure 3 for Just read twice: closing the recall gap for recurrent language models
Figure 4 for Just read twice: closing the recall gap for recurrent language models
Viaarxiv icon

State-Free Inference of State-Space Models: The Transfer Function Approach

Add code
May 10, 2024
Viaarxiv icon

Mechanistic Design and Scaling of Hybrid Architectures

Add code
Mar 26, 2024
Figure 1 for Mechanistic Design and Scaling of Hybrid Architectures
Figure 2 for Mechanistic Design and Scaling of Hybrid Architectures
Figure 3 for Mechanistic Design and Scaling of Hybrid Architectures
Figure 4 for Mechanistic Design and Scaling of Hybrid Architectures
Viaarxiv icon

Simple linear attention language models balance the recall-throughput tradeoff

Add code
Feb 28, 2024
Figure 1 for Simple linear attention language models balance the recall-throughput tradeoff
Figure 2 for Simple linear attention language models balance the recall-throughput tradeoff
Figure 3 for Simple linear attention language models balance the recall-throughput tradeoff
Figure 4 for Simple linear attention language models balance the recall-throughput tradeoff
Viaarxiv icon

Prospector Heads: Generalized Feature Attribution for Large Models & Data

Add code
Feb 18, 2024
Viaarxiv icon

Benchmarking and Building Long-Context Retrieval Models with LoCo and M2-BERT

Add code
Feb 14, 2024
Viaarxiv icon

Hydragen: High-Throughput LLM Inference with Shared Prefixes

Add code
Feb 07, 2024
Figure 1 for Hydragen: High-Throughput LLM Inference with Shared Prefixes
Figure 2 for Hydragen: High-Throughput LLM Inference with Shared Prefixes
Figure 3 for Hydragen: High-Throughput LLM Inference with Shared Prefixes
Figure 4 for Hydragen: High-Throughput LLM Inference with Shared Prefixes
Viaarxiv icon