Picture for Wenyong Zhou

Wenyong Zhou

Can We Trust LLMs on Memristors? Diving into Reasoning Ability under Non-Ideality

Add code
Mar 14, 2026
Viaarxiv icon

Binary Weight Multi-Bit Activation Quantization for Compute-in-Memory CNN Accelerators

Add code
Aug 29, 2025
Figure 1 for Binary Weight Multi-Bit Activation Quantization for Compute-in-Memory CNN Accelerators
Figure 2 for Binary Weight Multi-Bit Activation Quantization for Compute-in-Memory CNN Accelerators
Figure 3 for Binary Weight Multi-Bit Activation Quantization for Compute-in-Memory CNN Accelerators
Figure 4 for Binary Weight Multi-Bit Activation Quantization for Compute-in-Memory CNN Accelerators
Viaarxiv icon

QuadINR: Hardware-Efficient Implicit Neural Representations Through Quadratic Activation

Add code
Aug 20, 2025
Viaarxiv icon

HaLoRA: Hardware-aware Low-Rank Adaptation for Large Language Models Based on Hybrid Compute-in-Memory Architecture

Add code
Feb 27, 2025
Figure 1 for HaLoRA: Hardware-aware Low-Rank Adaptation for Large Language Models Based on Hybrid Compute-in-Memory Architecture
Figure 2 for HaLoRA: Hardware-aware Low-Rank Adaptation for Large Language Models Based on Hybrid Compute-in-Memory Architecture
Figure 3 for HaLoRA: Hardware-aware Low-Rank Adaptation for Large Language Models Based on Hybrid Compute-in-Memory Architecture
Figure 4 for HaLoRA: Hardware-aware Low-Rank Adaptation for Large Language Models Based on Hybrid Compute-in-Memory Architecture
Viaarxiv icon