Picture for Yanzhi Wang

Yanzhi Wang

Q-realign: Piggybacking Realignment on Quantization for Safe and Efficient LLM Deployment

Add code
Jan 13, 2026
Viaarxiv icon

From Bits to Chips: An LLM-based Hardware-Aware Quantization Agent for Streamlined Deployment of LLMs

Add code
Jan 07, 2026
Viaarxiv icon

Open-Source Multimodal Moxin Models with Moxin-VLM and Moxin-VLA

Add code
Dec 22, 2025
Viaarxiv icon

ZeroSim: Zero-Shot Analog Circuit Evaluation with Unified Transformer Embeddings

Add code
Nov 10, 2025
Viaarxiv icon

CircuitSense: A Hierarchical Circuit System Benchmark Bridging Visual Comprehension and Symbolic Reasoning in Engineering Design Process

Add code
Sep 26, 2025
Viaarxiv icon

TSLA: A Task-Specific Learning Adaptation for Semantic Segmentation on Autonomous Vehicles Platform

Add code
Aug 17, 2025
Figure 1 for TSLA: A Task-Specific Learning Adaptation for Semantic Segmentation on Autonomous Vehicles Platform
Figure 2 for TSLA: A Task-Specific Learning Adaptation for Semantic Segmentation on Autonomous Vehicles Platform
Figure 3 for TSLA: A Task-Specific Learning Adaptation for Semantic Segmentation on Autonomous Vehicles Platform
Figure 4 for TSLA: A Task-Specific Learning Adaptation for Semantic Segmentation on Autonomous Vehicles Platform
Viaarxiv icon

RCR-Router: Efficient Role-Aware Context Routing for Multi-Agent LLM Systems with Structured Memory

Add code
Aug 06, 2025
Viaarxiv icon

Taming Diffusion Transformer for Real-Time Mobile Video Generation

Add code
Jul 17, 2025
Figure 1 for Taming Diffusion Transformer for Real-Time Mobile Video Generation
Figure 2 for Taming Diffusion Transformer for Real-Time Mobile Video Generation
Figure 3 for Taming Diffusion Transformer for Real-Time Mobile Video Generation
Figure 4 for Taming Diffusion Transformer for Real-Time Mobile Video Generation
Viaarxiv icon

Enabling Flexible Multi-LLM Integration for Scalable Knowledge Aggregation

Add code
May 28, 2025
Figure 1 for Enabling Flexible Multi-LLM Integration for Scalable Knowledge Aggregation
Figure 2 for Enabling Flexible Multi-LLM Integration for Scalable Knowledge Aggregation
Figure 3 for Enabling Flexible Multi-LLM Integration for Scalable Knowledge Aggregation
Figure 4 for Enabling Flexible Multi-LLM Integration for Scalable Knowledge Aggregation
Viaarxiv icon

ALTER: All-in-One Layer Pruning and Temporal Expert Routing for Efficient Diffusion Generation

Add code
May 27, 2025
Viaarxiv icon