Picture for Chuan Wu

Chuan Wu

The University of Hong Kong

Laminar: A Scalable Asynchronous RL Post-Training Framework

Add code
Oct 14, 2025
Viaarxiv icon

On the Interplay between Graph Structure and Learning Algorithms in Graph Neural Networks

Add code
Aug 20, 2025
Viaarxiv icon

A Non-Asymptotic Convergent Analysis for Scored-Based Graph Generative Model via a System of Stochastic Differential Equations

Add code
Aug 20, 2025
Viaarxiv icon

OVERLORD: Ultimate Scaling of DataLoader for Multi-Source Large Foundation Model Training

Add code
Apr 14, 2025
Viaarxiv icon

TreeSynth: Synthesizing Diverse Data from Scratch via Tree-Guided Subspace Partitioning

Add code
Mar 21, 2025
Figure 1 for TreeSynth: Synthesizing Diverse Data from Scratch via Tree-Guided Subspace Partitioning
Figure 2 for TreeSynth: Synthesizing Diverse Data from Scratch via Tree-Guided Subspace Partitioning
Figure 3 for TreeSynth: Synthesizing Diverse Data from Scratch via Tree-Guided Subspace Partitioning
Figure 4 for TreeSynth: Synthesizing Diverse Data from Scratch via Tree-Guided Subspace Partitioning
Viaarxiv icon

Developing and Utilizing a Large-Scale Cantonese Dataset for Multi-Tasking in Large Language Models

Add code
Mar 05, 2025
Figure 1 for Developing and Utilizing a Large-Scale Cantonese Dataset for Multi-Tasking in Large Language Models
Figure 2 for Developing and Utilizing a Large-Scale Cantonese Dataset for Multi-Tasking in Large Language Models
Figure 3 for Developing and Utilizing a Large-Scale Cantonese Dataset for Multi-Tasking in Large Language Models
Figure 4 for Developing and Utilizing a Large-Scale Cantonese Dataset for Multi-Tasking in Large Language Models
Viaarxiv icon

Optimizing Distributed Deployment of Mixture-of-Experts Model Inference in Serverless Computing

Add code
Jan 09, 2025
Viaarxiv icon

Echo: Simulating Distributed Training At Scale

Add code
Dec 17, 2024
Figure 1 for Echo: Simulating Distributed Training At Scale
Figure 2 for Echo: Simulating Distributed Training At Scale
Figure 3 for Echo: Simulating Distributed Training At Scale
Figure 4 for Echo: Simulating Distributed Training At Scale
Viaarxiv icon

FedMoE-DA: Federated Mixture of Experts via Domain Aware Fine-grained Aggregation

Add code
Nov 04, 2024
Figure 1 for FedMoE-DA: Federated Mixture of Experts via Domain Aware Fine-grained Aggregation
Figure 2 for FedMoE-DA: Federated Mixture of Experts via Domain Aware Fine-grained Aggregation
Figure 3 for FedMoE-DA: Federated Mixture of Experts via Domain Aware Fine-grained Aggregation
Figure 4 for FedMoE-DA: Federated Mixture of Experts via Domain Aware Fine-grained Aggregation
Viaarxiv icon

Forewarned is Forearmed: Leveraging LLMs for Data Synthesis through Failure-Inducing Exploration

Add code
Oct 22, 2024
Figure 1 for Forewarned is Forearmed: Leveraging LLMs for Data Synthesis through Failure-Inducing Exploration
Figure 2 for Forewarned is Forearmed: Leveraging LLMs for Data Synthesis through Failure-Inducing Exploration
Figure 3 for Forewarned is Forearmed: Leveraging LLMs for Data Synthesis through Failure-Inducing Exploration
Figure 4 for Forewarned is Forearmed: Leveraging LLMs for Data Synthesis through Failure-Inducing Exploration
Viaarxiv icon