Picture for Minjia Zhang

Minjia Zhang

MedCite: Can Language Models Generate Verifiable Text for Medicine?

Add code
Jun 07, 2025
Viaarxiv icon

VTool-R1: VLMs Learn to Think with Images via Reinforcement Learning on Multimodal Tool Use

Add code
May 25, 2025
Viaarxiv icon

Teaching Large Language Models to Maintain Contextual Faithfulness via Synthetic Tasks and Reinforcement Learning

Add code
May 22, 2025
Viaarxiv icon

MiLo: Efficient Quantized MoE Inference with Mixture of Low-Rank Compensators

Add code
Apr 03, 2025
Viaarxiv icon

Cache-of-Thought: Master-Apprentice Framework for Cost-Effective Vision Language Model Inference

Add code
Feb 27, 2025
Viaarxiv icon

RAG-Gym: Optimizing Reasoning and Search Agents with Process Supervision

Add code
Feb 19, 2025
Viaarxiv icon

Next Token Prediction Towards Multimodal Intelligence: A Comprehensive Survey

Add code
Dec 30, 2024
Figure 1 for Next Token Prediction Towards Multimodal Intelligence: A Comprehensive Survey
Figure 2 for Next Token Prediction Towards Multimodal Intelligence: A Comprehensive Survey
Figure 3 for Next Token Prediction Towards Multimodal Intelligence: A Comprehensive Survey
Figure 4 for Next Token Prediction Towards Multimodal Intelligence: A Comprehensive Survey
Viaarxiv icon

MiniKV: Pushing the Limits of LLM Inference via 2-Bit Layer-Discriminative KV Cache

Add code
Nov 28, 2024
Figure 1 for MiniKV: Pushing the Limits of LLM Inference via 2-Bit Layer-Discriminative KV Cache
Figure 2 for MiniKV: Pushing the Limits of LLM Inference via 2-Bit Layer-Discriminative KV Cache
Figure 3 for MiniKV: Pushing the Limits of LLM Inference via 2-Bit Layer-Discriminative KV Cache
Figure 4 for MiniKV: Pushing the Limits of LLM Inference via 2-Bit Layer-Discriminative KV Cache
Viaarxiv icon

Pushing the Limits of LLM Inference via 2-Bit Layer-Discriminative KV Cache

Add code
Nov 27, 2024
Figure 1 for Pushing the Limits of LLM Inference via 2-Bit Layer-Discriminative KV Cache
Figure 2 for Pushing the Limits of LLM Inference via 2-Bit Layer-Discriminative KV Cache
Figure 3 for Pushing the Limits of LLM Inference via 2-Bit Layer-Discriminative KV Cache
Figure 4 for Pushing the Limits of LLM Inference via 2-Bit Layer-Discriminative KV Cache
Viaarxiv icon

Transforming the Hybrid Cloud for Emerging AI Workloads

Add code
Nov 20, 2024
Viaarxiv icon