Picture for Simran Arora

Simran Arora

ParallelKittens: Systematic and Practical Simplification of Multi-GPU AI Kernels

Add code
Nov 17, 2025
Viaarxiv icon

HipKittens: Fast and Furious AMD Kernels

Add code
Nov 11, 2025
Viaarxiv icon

Cartridges: Lightweight and general-purpose long context representations via self-study

Add code
Jun 06, 2025
Figure 1 for Cartridges: Lightweight and general-purpose long context representations via self-study
Figure 2 for Cartridges: Lightweight and general-purpose long context representations via self-study
Figure 3 for Cartridges: Lightweight and general-purpose long context representations via self-study
Figure 4 for Cartridges: Lightweight and general-purpose long context representations via self-study
Viaarxiv icon

Towards Learning High-Precision Least Squares Algorithms with Sequence Models

Add code
Mar 15, 2025
Figure 1 for Towards Learning High-Precision Least Squares Algorithms with Sequence Models
Figure 2 for Towards Learning High-Precision Least Squares Algorithms with Sequence Models
Figure 3 for Towards Learning High-Precision Least Squares Algorithms with Sequence Models
Figure 4 for Towards Learning High-Precision Least Squares Algorithms with Sequence Models
Viaarxiv icon

ThunderKittens: Simple, Fast, and Adorable AI Kernels

Add code
Oct 27, 2024
Viaarxiv icon

LoLCATs: On Low-Rank Linearizing of Large Language Models

Add code
Oct 14, 2024
Figure 1 for LoLCATs: On Low-Rank Linearizing of Large Language Models
Figure 2 for LoLCATs: On Low-Rank Linearizing of Large Language Models
Figure 3 for LoLCATs: On Low-Rank Linearizing of Large Language Models
Figure 4 for LoLCATs: On Low-Rank Linearizing of Large Language Models
Viaarxiv icon

Just read twice: closing the recall gap for recurrent language models

Add code
Jul 07, 2024
Figure 1 for Just read twice: closing the recall gap for recurrent language models
Figure 2 for Just read twice: closing the recall gap for recurrent language models
Figure 3 for Just read twice: closing the recall gap for recurrent language models
Figure 4 for Just read twice: closing the recall gap for recurrent language models
Viaarxiv icon

Optimistic Verifiable Training by Controlling Hardware Nondeterminism

Add code
Mar 16, 2024
Figure 1 for Optimistic Verifiable Training by Controlling Hardware Nondeterminism
Figure 2 for Optimistic Verifiable Training by Controlling Hardware Nondeterminism
Figure 3 for Optimistic Verifiable Training by Controlling Hardware Nondeterminism
Figure 4 for Optimistic Verifiable Training by Controlling Hardware Nondeterminism
Viaarxiv icon

Simple linear attention language models balance the recall-throughput tradeoff

Add code
Feb 28, 2024
Figure 1 for Simple linear attention language models balance the recall-throughput tradeoff
Figure 2 for Simple linear attention language models balance the recall-throughput tradeoff
Figure 3 for Simple linear attention language models balance the recall-throughput tradeoff
Figure 4 for Simple linear attention language models balance the recall-throughput tradeoff
Viaarxiv icon

Benchmarking and Building Long-Context Retrieval Models with LoCo and M2-BERT

Add code
Feb 14, 2024
Viaarxiv icon