Picture for Jason Cong

Jason Cong

LUT-LLM: Efficient Large Language Model Inference with Memory-based Computations on FPGAs

Add code
Nov 09, 2025
Viaarxiv icon

LLM-DSE: Searching Accelerator Parameters with LLM Agents

Add code
May 18, 2025
Figure 1 for LLM-DSE: Searching Accelerator Parameters with LLM Agents
Figure 2 for LLM-DSE: Searching Accelerator Parameters with LLM Agents
Figure 3 for LLM-DSE: Searching Accelerator Parameters with LLM Agents
Figure 4 for LLM-DSE: Searching Accelerator Parameters with LLM Agents
Viaarxiv icon

LIFT: LLM-Based Pragma Insertion for HLS via GNN Supervised Fine-Tuning

Add code
Apr 29, 2025
Figure 1 for LIFT: LLM-Based Pragma Insertion for HLS via GNN Supervised Fine-Tuning
Figure 2 for LIFT: LLM-Based Pragma Insertion for HLS via GNN Supervised Fine-Tuning
Figure 3 for LIFT: LLM-Based Pragma Insertion for HLS via GNN Supervised Fine-Tuning
Figure 4 for LIFT: LLM-Based Pragma Insertion for HLS via GNN Supervised Fine-Tuning
Viaarxiv icon

InTAR: Inter-Task Auto-Reconfigurable Accelerator Design for High Data Volume Variation in DNNs

Add code
Feb 12, 2025
Viaarxiv icon

Hierarchical Mixture of Experts: Generalizable Learning for High-Level Synthesis

Add code
Oct 25, 2024
Figure 1 for Hierarchical Mixture of Experts: Generalizable Learning for High-Level Synthesis
Figure 2 for Hierarchical Mixture of Experts: Generalizable Learning for High-Level Synthesis
Figure 3 for Hierarchical Mixture of Experts: Generalizable Learning for High-Level Synthesis
Figure 4 for Hierarchical Mixture of Experts: Generalizable Learning for High-Level Synthesis
Viaarxiv icon

Dynamic-Width Speculative Beam Decoding for Efficient LLM Inference

Add code
Sep 25, 2024
Figure 1 for Dynamic-Width Speculative Beam Decoding for Efficient LLM Inference
Figure 2 for Dynamic-Width Speculative Beam Decoding for Efficient LLM Inference
Figure 3 for Dynamic-Width Speculative Beam Decoding for Efficient LLM Inference
Figure 4 for Dynamic-Width Speculative Beam Decoding for Efficient LLM Inference
Viaarxiv icon

Accelerating Large Language Model Pretraining via LFR Pedagogy: Learn, Focus, and Review

Add code
Sep 10, 2024
Figure 1 for Accelerating Large Language Model Pretraining via LFR Pedagogy: Learn, Focus, and Review
Figure 2 for Accelerating Large Language Model Pretraining via LFR Pedagogy: Learn, Focus, and Review
Figure 3 for Accelerating Large Language Model Pretraining via LFR Pedagogy: Learn, Focus, and Review
Figure 4 for Accelerating Large Language Model Pretraining via LFR Pedagogy: Learn, Focus, and Review
Viaarxiv icon

Multi-Token Joint Speculative Decoding for Accelerating Large Language Model Inference

Add code
Jul 12, 2024
Figure 1 for Multi-Token Joint Speculative Decoding for Accelerating Large Language Model Inference
Figure 2 for Multi-Token Joint Speculative Decoding for Accelerating Large Language Model Inference
Figure 3 for Multi-Token Joint Speculative Decoding for Accelerating Large Language Model Inference
Figure 4 for Multi-Token Joint Speculative Decoding for Accelerating Large Language Model Inference
Viaarxiv icon

Cross-Modality Program Representation Learning for Electronic Design Automation with High-Level Synthesis

Add code
Jun 13, 2024
Viaarxiv icon

HMT: Hierarchical Memory Transformer for Long Context Language Processing

Add code
May 09, 2024
Viaarxiv icon