Picture for Vitaly Kurin

Vitaly Kurin

NVIDIA Nemotron 3: Efficient and Open Intelligence

Add code
Dec 24, 2025
Viaarxiv icon

Nemotron 3 Nano: Open, Efficient Mixture-of-Experts Hybrid Mamba-Transformer Model for Agentic Reasoning

Add code
Dec 23, 2025
Viaarxiv icon

NVIDIA Nemotron Nano 2: An Accurate and Efficient Hybrid Mamba-Transformer Reasoning Model

Add code
Aug 21, 2025
Figure 1 for NVIDIA Nemotron Nano 2: An Accurate and Efficient Hybrid Mamba-Transformer Reasoning Model
Figure 2 for NVIDIA Nemotron Nano 2: An Accurate and Efficient Hybrid Mamba-Transformer Reasoning Model
Figure 3 for NVIDIA Nemotron Nano 2: An Accurate and Efficient Hybrid Mamba-Transformer Reasoning Model
Figure 4 for NVIDIA Nemotron Nano 2: An Accurate and Efficient Hybrid Mamba-Transformer Reasoning Model
Viaarxiv icon

Nemotron-H: A Family of Accurate and Efficient Hybrid Mamba-Transformer Models

Add code
Apr 10, 2025
Figure 1 for Nemotron-H: A Family of Accurate and Efficient Hybrid Mamba-Transformer Models
Figure 2 for Nemotron-H: A Family of Accurate and Efficient Hybrid Mamba-Transformer Models
Figure 3 for Nemotron-H: A Family of Accurate and Efficient Hybrid Mamba-Transformer Models
Figure 4 for Nemotron-H: A Family of Accurate and Efficient Hybrid Mamba-Transformer Models
Viaarxiv icon

A Generalist Neural Algorithmic Learner

Add code
Sep 22, 2022
Figure 1 for A Generalist Neural Algorithmic Learner
Figure 2 for A Generalist Neural Algorithmic Learner
Figure 3 for A Generalist Neural Algorithmic Learner
Figure 4 for A Generalist Neural Algorithmic Learner
Viaarxiv icon

Insights From the NeurIPS 2021 NetHack Challenge

Add code
Mar 22, 2022
Figure 1 for Insights From the NeurIPS 2021 NetHack Challenge
Figure 2 for Insights From the NeurIPS 2021 NetHack Challenge
Figure 3 for Insights From the NeurIPS 2021 NetHack Challenge
Figure 4 for Insights From the NeurIPS 2021 NetHack Challenge
Viaarxiv icon

You May Not Need Ratio Clipping in PPO

Add code
Jan 31, 2022
Figure 1 for You May Not Need Ratio Clipping in PPO
Figure 2 for You May Not Need Ratio Clipping in PPO
Figure 3 for You May Not Need Ratio Clipping in PPO
Figure 4 for You May Not Need Ratio Clipping in PPO
Viaarxiv icon

In Defense of the Unitary Scalarization for Deep Multi-Task Learning

Add code
Jan 20, 2022
Figure 1 for In Defense of the Unitary Scalarization for Deep Multi-Task Learning
Figure 2 for In Defense of the Unitary Scalarization for Deep Multi-Task Learning
Figure 3 for In Defense of the Unitary Scalarization for Deep Multi-Task Learning
Figure 4 for In Defense of the Unitary Scalarization for Deep Multi-Task Learning
Viaarxiv icon

MiniHack the Planet: A Sandbox for Open-Ended Reinforcement Learning Research

Add code
Sep 27, 2021
Figure 1 for MiniHack the Planet: A Sandbox for Open-Ended Reinforcement Learning Research
Figure 2 for MiniHack the Planet: A Sandbox for Open-Ended Reinforcement Learning Research
Figure 3 for MiniHack the Planet: A Sandbox for Open-Ended Reinforcement Learning Research
Figure 4 for MiniHack the Planet: A Sandbox for Open-Ended Reinforcement Learning Research
Viaarxiv icon

Snowflake: Scaling GNNs to High-Dimensional Continuous Control via Parameter Freezing

Add code
Mar 01, 2021
Figure 1 for Snowflake: Scaling GNNs to High-Dimensional Continuous Control via Parameter Freezing
Figure 2 for Snowflake: Scaling GNNs to High-Dimensional Continuous Control via Parameter Freezing
Figure 3 for Snowflake: Scaling GNNs to High-Dimensional Continuous Control via Parameter Freezing
Figure 4 for Snowflake: Scaling GNNs to High-Dimensional Continuous Control via Parameter Freezing
Viaarxiv icon