Picture for Zhiru Zhang

Zhiru Zhang

Radial Networks: Dynamic Layer Routing for High-Performance Large Language Models

Add code
Apr 07, 2024
Viaarxiv icon

Allo: A Programming Model for Composable Accelerator Design

Add code
Apr 07, 2024
Figure 1 for Allo: A Programming Model for Composable Accelerator Design
Figure 2 for Allo: A Programming Model for Composable Accelerator Design
Figure 3 for Allo: A Programming Model for Composable Accelerator Design
Figure 4 for Allo: A Programming Model for Composable Accelerator Design
Viaarxiv icon

UniSparse: An Intermediate Language for General Sparse Format Customization

Add code
Mar 09, 2024
Figure 1 for UniSparse: An Intermediate Language for General Sparse Format Customization
Figure 2 for UniSparse: An Intermediate Language for General Sparse Format Customization
Figure 3 for UniSparse: An Intermediate Language for General Sparse Format Customization
Figure 4 for UniSparse: An Intermediate Language for General Sparse Format Customization
Viaarxiv icon

Less is More: Hop-Wise Graph Attention for Scalable and Generalizable Learning on Circuits

Add code
Mar 06, 2024
Viaarxiv icon

Polynormer: Polynomial-Expressive Graph Transformer in Linear Time

Add code
Mar 02, 2024
Figure 1 for Polynormer: Polynomial-Expressive Graph Transformer in Linear Time
Figure 2 for Polynormer: Polynomial-Expressive Graph Transformer in Linear Time
Figure 3 for Polynormer: Polynomial-Expressive Graph Transformer in Linear Time
Figure 4 for Polynormer: Polynomial-Expressive Graph Transformer in Linear Time
Viaarxiv icon

Exploring the Limits of Semantic Image Compression at Micro-bits per Pixel

Add code
Feb 21, 2024
Figure 1 for Exploring the Limits of Semantic Image Compression at Micro-bits per Pixel
Figure 2 for Exploring the Limits of Semantic Image Compression at Micro-bits per Pixel
Figure 3 for Exploring the Limits of Semantic Image Compression at Micro-bits per Pixel
Figure 4 for Exploring the Limits of Semantic Image Compression at Micro-bits per Pixel
Viaarxiv icon

SAGMAN: Stability Analysis of Graph Neural Networks on the Manifolds

Add code
Feb 21, 2024
Viaarxiv icon

Trainable Fixed-Point Quantization for Deep Learning Acceleration on FPGAs

Add code
Jan 31, 2024
Viaarxiv icon

Understanding the Potential of FPGA-Based Spatial Acceleration for Large Language Model Inference

Add code
Dec 23, 2023
Figure 1 for Understanding the Potential of FPGA-Based Spatial Acceleration for Large Language Model Inference
Figure 2 for Understanding the Potential of FPGA-Based Spatial Acceleration for Large Language Model Inference
Figure 3 for Understanding the Potential of FPGA-Based Spatial Acceleration for Large Language Model Inference
Figure 4 for Understanding the Potential of FPGA-Based Spatial Acceleration for Large Language Model Inference
Viaarxiv icon

FLIQS: One-Shot Mixed-Precision Floating-Point and Integer Quantization Search

Add code
Aug 07, 2023
Figure 1 for FLIQS: One-Shot Mixed-Precision Floating-Point and Integer Quantization Search
Figure 2 for FLIQS: One-Shot Mixed-Precision Floating-Point and Integer Quantization Search
Figure 3 for FLIQS: One-Shot Mixed-Precision Floating-Point and Integer Quantization Search
Figure 4 for FLIQS: One-Shot Mixed-Precision Floating-Point and Integer Quantization Search
Viaarxiv icon