Picture for Zhiheng Cheng

Zhiheng Cheng

TeLLMe: An Energy-Efficient Ternary LLM Accelerator for Prefilling and Decoding on Edge FPGAs

Add code
Apr 22, 2025
Viaarxiv icon

COBRA: Algorithm-Architecture Co-optimized Binary Transformer Accelerator for Edge Inference

Add code
Apr 22, 2025
Figure 1 for COBRA: Algorithm-Architecture Co-optimized Binary Transformer Accelerator for Edge Inference
Figure 2 for COBRA: Algorithm-Architecture Co-optimized Binary Transformer Accelerator for Edge Inference
Figure 3 for COBRA: Algorithm-Architecture Co-optimized Binary Transformer Accelerator for Edge Inference
Figure 4 for COBRA: Algorithm-Architecture Co-optimized Binary Transformer Accelerator for Edge Inference
Viaarxiv icon

Unleashing the Potential of SAM for Medical Adaptation via Hierarchical Decoding

Add code
Mar 27, 2024
Viaarxiv icon