Picture for Fangming Liu

Fangming Liu

Hybrid-Parallel: Achieving High Performance and Energy Efficient Distributed Inference on Robots

Add code
May 29, 2024
Viaarxiv icon

TrimCaching: Parameter-sharing AI Model Caching in Wireless Edge Networks

Add code
May 07, 2024
Viaarxiv icon

Opara: Exploiting Operator Parallelism for Expediting DNN Inference on GPUs

Add code
Dec 16, 2023
Figure 1 for Opara: Exploiting Operator Parallelism for Expediting DNN Inference on GPUs
Figure 2 for Opara: Exploiting Operator Parallelism for Expediting DNN Inference on GPUs
Figure 3 for Opara: Exploiting Operator Parallelism for Expediting DNN Inference on GPUs
Figure 4 for Opara: Exploiting Operator Parallelism for Expediting DNN Inference on GPUs
Viaarxiv icon

On-edge Multi-task Transfer Learning: Model and Practice with Data-driven Task Allocation

Add code
Jul 06, 2021
Figure 1 for On-edge Multi-task Transfer Learning: Model and Practice with Data-driven Task Allocation
Figure 2 for On-edge Multi-task Transfer Learning: Model and Practice with Data-driven Task Allocation
Figure 3 for On-edge Multi-task Transfer Learning: Model and Practice with Data-driven Task Allocation
Figure 4 for On-edge Multi-task Transfer Learning: Model and Practice with Data-driven Task Allocation
Viaarxiv icon