Picture for Lili Qiu

Lili Qiu

RetrievalAttention: Accelerating Long-Context LLM Inference via Vector Retrieval

Add code
Sep 16, 2024
Viaarxiv icon

Advancing Multi-Modal Sensing Through Expandable Modality Alignment

Add code
Jul 25, 2024
Viaarxiv icon

MInference 1.0: Accelerating Pre-filling for Long-Context LLMs via Dynamic Sparse Attention

Add code
Jul 02, 2024
Viaarxiv icon

Expressive and Generalizable Low-rank Adaptation for Large Models via Slow Cascaded Learning

Add code
Jul 01, 2024
Viaarxiv icon

Mitigate Position Bias in Large Language Models via Scaling a Single Dimension

Add code
Jun 04, 2024
Viaarxiv icon

Parrot: Efficient Serving of LLM-based Applications with Semantic Variable

Add code
May 30, 2024
Viaarxiv icon

Position Engineering: Boosting Large Language Models through Positional Information Manipulation

Add code
Apr 17, 2024
Viaarxiv icon

LLM-ABR: Designing Adaptive Bitrate Algorithms via Large Language Models

Add code
Apr 02, 2024
Viaarxiv icon

LLM-RadJudge: Achieving Radiologist-Level Evaluation for X-Ray Report Generation

Add code
Apr 01, 2024
Viaarxiv icon

LLMLingua-2: Data Distillation for Efficient and Faithful Task-Agnostic Prompt Compression

Add code
Mar 19, 2024
Figure 1 for LLMLingua-2: Data Distillation for Efficient and Faithful Task-Agnostic Prompt Compression
Figure 2 for LLMLingua-2: Data Distillation for Efficient and Faithful Task-Agnostic Prompt Compression
Figure 3 for LLMLingua-2: Data Distillation for Efficient and Faithful Task-Agnostic Prompt Compression
Figure 4 for LLMLingua-2: Data Distillation for Efficient and Faithful Task-Agnostic Prompt Compression
Viaarxiv icon