Picture for Samuel Kriman

Samuel Kriman

NVIDIA Nemotron Nano 2: An Accurate and Efficient Hybrid Mamba-Transformer Reasoning Model

Add code
Aug 21, 2025
Viaarxiv icon

SWAN-GPT: An Efficient and Scalable Approach for Long-Context Language Modeling

Add code
Apr 11, 2025
Viaarxiv icon

Nemotron-H: A Family of Accurate and Efficient Hybrid Mamba-Transformer Models

Add code
Apr 10, 2025
Viaarxiv icon

RULER: What's the Real Context Size of Your Long-Context Language Models?

Add code
Apr 11, 2024
Viaarxiv icon

Investigating End-to-End ASR Architectures for Long Form Audio Transcription

Add code
Sep 20, 2023
Figure 1 for Investigating End-to-End ASR Architectures for Long Form Audio Transcription
Figure 2 for Investigating End-to-End ASR Architectures for Long Form Audio Transcription
Figure 3 for Investigating End-to-End ASR Architectures for Long Form Audio Transcription
Figure 4 for Investigating End-to-End ASR Architectures for Long Form Audio Transcription
Viaarxiv icon

Fast Conformer with Linearly Scalable Attention for Efficient Speech Recognition

Add code
May 19, 2023
Figure 1 for Fast Conformer with Linearly Scalable Attention for Efficient Speech Recognition
Figure 2 for Fast Conformer with Linearly Scalable Attention for Efficient Speech Recognition
Figure 3 for Fast Conformer with Linearly Scalable Attention for Efficient Speech Recognition
Figure 4 for Fast Conformer with Linearly Scalable Attention for Efficient Speech Recognition
Viaarxiv icon

Accidental Learners: Spoken Language Identification in Multilingual Self-Supervised Models

Add code
Nov 09, 2022
Viaarxiv icon

NeMo: a toolkit for building AI applications using Neural Modules

Add code
Sep 14, 2019
Figure 1 for NeMo: a toolkit for building AI applications using Neural Modules
Figure 2 for NeMo: a toolkit for building AI applications using Neural Modules
Viaarxiv icon