Picture for Zhihua Wu

Zhihua Wu

StickMotion: Generating 3D Human Motions by Drawing a Stickman

Add code
Mar 05, 2025
Viaarxiv icon

ChuXin: 1.6B Technical Report

Add code
May 08, 2024
Figure 1 for ChuXin: 1.6B Technical Report
Figure 2 for ChuXin: 1.6B Technical Report
Figure 3 for ChuXin: 1.6B Technical Report
Figure 4 for ChuXin: 1.6B Technical Report
Viaarxiv icon

Efficient LLM Inference with Kcache

Add code
Apr 28, 2024
Viaarxiv icon

Code Comparison Tuning for Code Large Language Models

Add code
Mar 28, 2024
Figure 1 for Code Comparison Tuning for Code Large Language Models
Figure 2 for Code Comparison Tuning for Code Large Language Models
Figure 3 for Code Comparison Tuning for Code Large Language Models
Figure 4 for Code Comparison Tuning for Code Large Language Models
Viaarxiv icon

RecycleGPT: An Autoregressive Language Model with Recyclable Module

Add code
Aug 08, 2023
Viaarxiv icon

TA-MoE: Topology-Aware Large Scale Mixture-of-Expert Training

Add code
Feb 20, 2023
Viaarxiv icon

HelixFold: An Efficient Implementation of AlphaFold2 using PaddlePaddle

Add code
Jul 13, 2022
Figure 1 for HelixFold: An Efficient Implementation of AlphaFold2 using PaddlePaddle
Figure 2 for HelixFold: An Efficient Implementation of AlphaFold2 using PaddlePaddle
Figure 3 for HelixFold: An Efficient Implementation of AlphaFold2 using PaddlePaddle
Figure 4 for HelixFold: An Efficient Implementation of AlphaFold2 using PaddlePaddle
Viaarxiv icon

SE-MoE: A Scalable and Efficient Mixture-of-Experts Distributed Training and Inference System

Add code
May 20, 2022
Figure 1 for SE-MoE: A Scalable and Efficient Mixture-of-Experts Distributed Training and Inference System
Figure 2 for SE-MoE: A Scalable and Efficient Mixture-of-Experts Distributed Training and Inference System
Figure 3 for SE-MoE: A Scalable and Efficient Mixture-of-Experts Distributed Training and Inference System
Figure 4 for SE-MoE: A Scalable and Efficient Mixture-of-Experts Distributed Training and Inference System
Viaarxiv icon

Nebula-I: A General Framework for Collaboratively Training Deep Learning Models on Low-Bandwidth Cloud Clusters

Add code
May 19, 2022
Figure 1 for Nebula-I: A General Framework for Collaboratively Training Deep Learning Models on Low-Bandwidth Cloud Clusters
Figure 2 for Nebula-I: A General Framework for Collaboratively Training Deep Learning Models on Low-Bandwidth Cloud Clusters
Figure 3 for Nebula-I: A General Framework for Collaboratively Training Deep Learning Models on Low-Bandwidth Cloud Clusters
Figure 4 for Nebula-I: A General Framework for Collaboratively Training Deep Learning Models on Low-Bandwidth Cloud Clusters
Viaarxiv icon

ERNIE-ViLG: Unified Generative Pre-training for Bidirectional Vision-Language Generation

Add code
Dec 31, 2021
Figure 1 for ERNIE-ViLG: Unified Generative Pre-training for Bidirectional Vision-Language Generation
Figure 2 for ERNIE-ViLG: Unified Generative Pre-training for Bidirectional Vision-Language Generation
Figure 3 for ERNIE-ViLG: Unified Generative Pre-training for Bidirectional Vision-Language Generation
Figure 4 for ERNIE-ViLG: Unified Generative Pre-training for Bidirectional Vision-Language Generation
Viaarxiv icon