Picture for Christopher Ré

Christopher Ré

Department of Computer Science, Stanford University

Just read twice: closing the recall gap for recurrent language models

Add code
Jul 07, 2024
Viaarxiv icon

State-Free Inference of State-Space Models: The Transfer Function Approach

Add code
May 10, 2024
Viaarxiv icon

Mechanistic Design and Scaling of Hybrid Architectures

Add code
Mar 26, 2024
Figure 1 for Mechanistic Design and Scaling of Hybrid Architectures
Figure 2 for Mechanistic Design and Scaling of Hybrid Architectures
Figure 3 for Mechanistic Design and Scaling of Hybrid Architectures
Figure 4 for Mechanistic Design and Scaling of Hybrid Architectures
Viaarxiv icon

Simple linear attention language models balance the recall-throughput tradeoff

Add code
Feb 28, 2024
Figure 1 for Simple linear attention language models balance the recall-throughput tradeoff
Figure 2 for Simple linear attention language models balance the recall-throughput tradeoff
Figure 3 for Simple linear attention language models balance the recall-throughput tradeoff
Figure 4 for Simple linear attention language models balance the recall-throughput tradeoff
Viaarxiv icon

Prospector Heads: Generalized Feature Attribution for Large Models & Data

Add code
Feb 18, 2024
Figure 1 for Prospector Heads: Generalized Feature Attribution for Large Models & Data
Figure 2 for Prospector Heads: Generalized Feature Attribution for Large Models & Data
Figure 3 for Prospector Heads: Generalized Feature Attribution for Large Models & Data
Figure 4 for Prospector Heads: Generalized Feature Attribution for Large Models & Data
Viaarxiv icon

Benchmarking and Building Long-Context Retrieval Models with LoCo and M2-BERT

Add code
Feb 14, 2024
Figure 1 for Benchmarking and Building Long-Context Retrieval Models with LoCo and M2-BERT
Figure 2 for Benchmarking and Building Long-Context Retrieval Models with LoCo and M2-BERT
Figure 3 for Benchmarking and Building Long-Context Retrieval Models with LoCo and M2-BERT
Figure 4 for Benchmarking and Building Long-Context Retrieval Models with LoCo and M2-BERT
Viaarxiv icon

Hydragen: High-Throughput LLM Inference with Shared Prefixes

Add code
Feb 07, 2024
Figure 1 for Hydragen: High-Throughput LLM Inference with Shared Prefixes
Figure 2 for Hydragen: High-Throughput LLM Inference with Shared Prefixes
Figure 3 for Hydragen: High-Throughput LLM Inference with Shared Prefixes
Figure 4 for Hydragen: High-Throughput LLM Inference with Shared Prefixes
Viaarxiv icon

The Hedgehog & the Porcupine: Expressive Linear Attentions with Softmax Mimicry

Add code
Feb 06, 2024
Figure 1 for The Hedgehog & the Porcupine: Expressive Linear Attentions with Softmax Mimicry
Figure 2 for The Hedgehog & the Porcupine: Expressive Linear Attentions with Softmax Mimicry
Figure 3 for The Hedgehog & the Porcupine: Expressive Linear Attentions with Softmax Mimicry
Figure 4 for The Hedgehog & the Porcupine: Expressive Linear Attentions with Softmax Mimicry
Viaarxiv icon

Zoology: Measuring and Improving Recall in Efficient Language Models

Add code
Dec 08, 2023
Figure 1 for Zoology: Measuring and Improving Recall in Efficient Language Models
Figure 2 for Zoology: Measuring and Improving Recall in Efficient Language Models
Figure 3 for Zoology: Measuring and Improving Recall in Efficient Language Models
Figure 4 for Zoology: Measuring and Improving Recall in Efficient Language Models
Viaarxiv icon

FlashFFTConv: Efficient Convolutions for Long Sequences with Tensor Cores

Add code
Nov 10, 2023
Figure 1 for FlashFFTConv: Efficient Convolutions for Long Sequences with Tensor Cores
Figure 2 for FlashFFTConv: Efficient Convolutions for Long Sequences with Tensor Cores
Figure 3 for FlashFFTConv: Efficient Convolutions for Long Sequences with Tensor Cores
Figure 4 for FlashFFTConv: Efficient Convolutions for Long Sequences with Tensor Cores
Viaarxiv icon