Picture for Minjia Zhang

Minjia Zhang

Model Tells You Where to Merge: Adaptive KV Cache Merging for LLMs on Long-Context Tasks

Add code
Jul 11, 2024
Viaarxiv icon

UltraEdit: Instruction-based Fine-Grained Image Editing at Scale

Add code
Jul 07, 2024
Figure 1 for UltraEdit: Instruction-based Fine-Grained Image Editing at Scale
Figure 2 for UltraEdit: Instruction-based Fine-Grained Image Editing at Scale
Figure 3 for UltraEdit: Instruction-based Fine-Grained Image Editing at Scale
Figure 4 for UltraEdit: Instruction-based Fine-Grained Image Editing at Scale
Viaarxiv icon

Universal Checkpointing: Efficient and Flexible Checkpointing for Large Scale Distributed Training

Add code
Jun 27, 2024
Viaarxiv icon

Computing in the Era of Large Generative Models: From Cloud-Native to AI-Native

Add code
Jan 17, 2024
Figure 1 for Computing in the Era of Large Generative Models: From Cloud-Native to AI-Native
Figure 2 for Computing in the Era of Large Generative Models: From Cloud-Native to AI-Native
Viaarxiv icon

DeepSpeed4Science Initiative: Enabling Large-Scale Scientific Discovery through Sophisticated AI System Technologies

Add code
Oct 11, 2023
Figure 1 for DeepSpeed4Science Initiative: Enabling Large-Scale Scientific Discovery through Sophisticated AI System Technologies
Figure 2 for DeepSpeed4Science Initiative: Enabling Large-Scale Scientific Discovery through Sophisticated AI System Technologies
Figure 3 for DeepSpeed4Science Initiative: Enabling Large-Scale Scientific Discovery through Sophisticated AI System Technologies
Figure 4 for DeepSpeed4Science Initiative: Enabling Large-Scale Scientific Discovery through Sophisticated AI System Technologies
Viaarxiv icon

Model Tells You What to Discard: Adaptive KV Cache Compression for LLMs

Add code
Oct 07, 2023
Figure 1 for Model Tells You What to Discard: Adaptive KV Cache Compression for LLMs
Figure 2 for Model Tells You What to Discard: Adaptive KV Cache Compression for LLMs
Figure 3 for Model Tells You What to Discard: Adaptive KV Cache Compression for LLMs
Figure 4 for Model Tells You What to Discard: Adaptive KV Cache Compression for LLMs
Viaarxiv icon

DeepSpeed-VisualChat: Multi-Round Multi-Image Interleave Chat via Multi-Modal Causal Attention

Add code
Sep 29, 2023
Viaarxiv icon

DeepSpeed Ulysses: System Optimizations for Enabling Training of Extreme Long Sequence Transformer Models

Add code
Sep 25, 2023
Viaarxiv icon

RenAIssance: A Survey into AI Text-to-Image Generation in the Era of Large Model

Add code
Sep 02, 2023
Figure 1 for RenAIssance: A Survey into AI Text-to-Image Generation in the Era of Large Model
Figure 2 for RenAIssance: A Survey into AI Text-to-Image Generation in the Era of Large Model
Figure 3 for RenAIssance: A Survey into AI Text-to-Image Generation in the Era of Large Model
Figure 4 for RenAIssance: A Survey into AI Text-to-Image Generation in the Era of Large Model
Viaarxiv icon

Cost-effective On-device Continual Learning over Memory Hierarchy with Miro

Add code
Aug 17, 2023
Figure 1 for Cost-effective On-device Continual Learning over Memory Hierarchy with Miro
Figure 2 for Cost-effective On-device Continual Learning over Memory Hierarchy with Miro
Figure 3 for Cost-effective On-device Continual Learning over Memory Hierarchy with Miro
Figure 4 for Cost-effective On-device Continual Learning over Memory Hierarchy with Miro
Viaarxiv icon