Picture for Haibo Chen

Haibo Chen

SmallThinker: A Family of Efficient Large Language Models Natively Trained for Local Deployment

Add code
Jul 28, 2025
Viaarxiv icon

An AI-native experimental laboratory for autonomous biomolecular engineering

Add code
Jul 03, 2025
Viaarxiv icon

Towards Multi-modal Graph Large Language Model

Add code
Jun 11, 2025
Viaarxiv icon

Get Experience from Practice: LLM Agents with Record & Replay

Add code
May 23, 2025
Viaarxiv icon

Modular Machine Learning: An Indispensable Path towards New-Generation Large Language Models

Add code
Apr 28, 2025
Figure 1 for Modular Machine Learning: An Indispensable Path towards New-Generation Large Language Models
Figure 2 for Modular Machine Learning: An Indispensable Path towards New-Generation Large Language Models
Figure 3 for Modular Machine Learning: An Indispensable Path towards New-Generation Large Language Models
Viaarxiv icon

L3: DIMM-PIM Integrated Architecture and Coordination for Scalable Long-Context LLM Inference

Add code
Apr 24, 2025
Viaarxiv icon

PipeWeaver: Addressing Data Dynamicity in Large Multimodal Model Training with Dynamic Interleaved Pipeline

Add code
Apr 19, 2025
Figure 1 for PipeWeaver: Addressing Data Dynamicity in Large Multimodal Model Training with Dynamic Interleaved Pipeline
Figure 2 for PipeWeaver: Addressing Data Dynamicity in Large Multimodal Model Training with Dynamic Interleaved Pipeline
Figure 3 for PipeWeaver: Addressing Data Dynamicity in Large Multimodal Model Training with Dynamic Interleaved Pipeline
Figure 4 for PipeWeaver: Addressing Data Dynamicity in Large Multimodal Model Training with Dynamic Interleaved Pipeline
Viaarxiv icon

AttentionEngine: A Versatile Framework for Efficient Attention Mechanisms on Diverse Hardware Platforms

Add code
Feb 21, 2025
Viaarxiv icon

Accelerating Large Language Models through Partially Linear Feed-Forward Network

Add code
Jan 17, 2025
Figure 1 for Accelerating Large Language Models through Partially Linear Feed-Forward Network
Figure 2 for Accelerating Large Language Models through Partially Linear Feed-Forward Network
Figure 3 for Accelerating Large Language Models through Partially Linear Feed-Forward Network
Figure 4 for Accelerating Large Language Models through Partially Linear Feed-Forward Network
Viaarxiv icon

KunServe: Elastic and Efficient Large Language Model Serving with Parameter-centric Memory Management

Add code
Dec 24, 2024
Viaarxiv icon