Picture for Sanjiv Kumar

Sanjiv Kumar

Google Research

SOAR: Improved Indexing for Approximate Nearest Neighbor Search

Add code
Mar 31, 2024
Viaarxiv icon

Metric-aware LLM inference

Add code
Mar 07, 2024
Viaarxiv icon

HiRE: High Recall Approximate Top-$k$ Estimation for Efficient LLM Inference

Add code
Feb 14, 2024
Figure 1 for HiRE: High Recall Approximate Top-$k$ Estimation for Efficient LLM Inference
Figure 2 for HiRE: High Recall Approximate Top-$k$ Estimation for Efficient LLM Inference
Figure 3 for HiRE: High Recall Approximate Top-$k$ Estimation for Efficient LLM Inference
Figure 4 for HiRE: High Recall Approximate Top-$k$ Estimation for Efficient LLM Inference
Viaarxiv icon

Tandem Transformers for Inference Efficient LLMs

Add code
Feb 13, 2024
Figure 1 for Tandem Transformers for Inference Efficient LLMs
Figure 2 for Tandem Transformers for Inference Efficient LLMs
Figure 3 for Tandem Transformers for Inference Efficient LLMs
Figure 4 for Tandem Transformers for Inference Efficient LLMs
Viaarxiv icon

Efficient Stagewise Pretraining via Progressive Subnetworks

Add code
Feb 08, 2024
Viaarxiv icon

SpacTor-T5: Pre-training T5 Models with Span Corruption and Replaced Token Detection

Add code
Jan 24, 2024
Viaarxiv icon

A Weighted K-Center Algorithm for Data Subset Selection

Add code
Dec 17, 2023
Viaarxiv icon

ReST meets ReAct: Self-Improvement for Multi-Step Reasoning LLM Agent

Add code
Dec 15, 2023
Viaarxiv icon

It's an Alignment, Not a Trade-off: Revisiting Bias and Variance in Deep Models

Add code
Oct 13, 2023
Viaarxiv icon

DistillSpec: Improving Speculative Decoding via Knowledge Distillation

Add code
Oct 12, 2023
Figure 1 for DistillSpec: Improving Speculative Decoding via Knowledge Distillation
Figure 2 for DistillSpec: Improving Speculative Decoding via Knowledge Distillation
Figure 3 for DistillSpec: Improving Speculative Decoding via Knowledge Distillation
Figure 4 for DistillSpec: Improving Speculative Decoding via Knowledge Distillation
Viaarxiv icon