Picture for Krishna Teja Chitty-Venkata

Krishna Teja Chitty-Venkata

PagedEviction: Structured Block-wise KV Cache Pruning for Efficient Large Language Model Inference

Add code
Sep 04, 2025
Viaarxiv icon

MoE-Inference-Bench: Performance Evaluation of Mixture of Expert Large Language and Vision Models

Add code
Aug 24, 2025
Viaarxiv icon

LangVision-LoRA-NAS: Neural Architecture Search for Variable LoRA Rank in Vision Language Models

Add code
Aug 17, 2025
Viaarxiv icon

BaKlaVa -- Budgeted Allocation of KV cache for Long-context Inference

Add code
Feb 18, 2025
Viaarxiv icon

LLM-Inference-Bench: Inference Benchmarking of Large Language Models on AI Accelerators

Add code
Oct 31, 2024
Viaarxiv icon

A Survey of Techniques for Optimizing Transformer Inference

Add code
Jul 16, 2023
Figure 1 for A Survey of Techniques for Optimizing Transformer Inference
Figure 2 for A Survey of Techniques for Optimizing Transformer Inference
Figure 3 for A Survey of Techniques for Optimizing Transformer Inference
Figure 4 for A Survey of Techniques for Optimizing Transformer Inference
Viaarxiv icon