Picture for Pavlo Molchanov

Pavlo Molchanov

$R^2$-dLLM: Accelerating Diffusion Large Language Models via Spatio-Temporal Redundancy Reduction

Add code
Apr 21, 2026
Viaarxiv icon

Nemotron 3 Super: Open, Efficient Mixture-of-Experts Hybrid Mamba-Transformer Model for Agentic Reasoning

Add code
Apr 14, 2026
Viaarxiv icon

Fast-dVLM: Efficient Block-Diffusion VLM via Direct Conversion from Autoregressive VLM

Add code
Apr 08, 2026
Viaarxiv icon

Attend Before Attention: Efficient and Scalable Video Understanding via Autoregressive Gazing

Add code
Mar 12, 2026
Viaarxiv icon

Stateful Token Reduction for Long-Video Hybrid VLMs

Add code
Feb 27, 2026
Viaarxiv icon

Extending Puzzle for Mixture-of-Experts Reasoning Models with Application to GPT-OSS Acceleration

Add code
Feb 12, 2026
Viaarxiv icon

C-RADIOv4 (Tech Report)

Add code
Jan 24, 2026
Viaarxiv icon

NVIDIA Nemotron 3: Efficient and Open Intelligence

Add code
Dec 24, 2025
Viaarxiv icon

Nemotron 3 Nano: Open, Efficient Mixture-of-Experts Hybrid Mamba-Transformer Model for Agentic Reasoning

Add code
Dec 23, 2025
Viaarxiv icon

Efficient-DLM: From Autoregressive to Diffusion Language Models, and Beyond in Speed

Add code
Dec 16, 2025
Figure 1 for Efficient-DLM: From Autoregressive to Diffusion Language Models, and Beyond in Speed
Figure 2 for Efficient-DLM: From Autoregressive to Diffusion Language Models, and Beyond in Speed
Figure 3 for Efficient-DLM: From Autoregressive to Diffusion Language Models, and Beyond in Speed
Figure 4 for Efficient-DLM: From Autoregressive to Diffusion Language Models, and Beyond in Speed
Viaarxiv icon