Picture for Dima Rekesh

Dima Rekesh

NVIDIA Nemotron Nano 2: An Accurate and Efficient Hybrid Mamba-Transformer Reasoning Model

Add code
Aug 21, 2025
Viaarxiv icon

Llama-Nemotron: Efficient Reasoning Models

Add code
May 02, 2025
Viaarxiv icon

SWAN-GPT: An Efficient and Scalable Approach for Long-Context Language Modeling

Add code
Apr 11, 2025
Viaarxiv icon

Nemotron-H: A Family of Accurate and Efficient Hybrid Mamba-Transformer Models

Add code
Apr 10, 2025
Viaarxiv icon

RULER: What's the Real Context Size of Your Long-Context Language Models?

Add code
Apr 11, 2024
Viaarxiv icon

Investigating End-to-End ASR Architectures for Long Form Audio Transcription

Add code
Sep 20, 2023
Figure 1 for Investigating End-to-End ASR Architectures for Long Form Audio Transcription
Figure 2 for Investigating End-to-End ASR Architectures for Long Form Audio Transcription
Figure 3 for Investigating End-to-End ASR Architectures for Long Form Audio Transcription
Figure 4 for Investigating End-to-End ASR Architectures for Long Form Audio Transcription
Viaarxiv icon

Towards training Bilingual and Code-Switched Speech Recognition models from Monolingual data sources

Add code
Jun 14, 2023
Figure 1 for Towards training Bilingual and Code-Switched Speech Recognition models from Monolingual data sources
Figure 2 for Towards training Bilingual and Code-Switched Speech Recognition models from Monolingual data sources
Figure 3 for Towards training Bilingual and Code-Switched Speech Recognition models from Monolingual data sources
Figure 4 for Towards training Bilingual and Code-Switched Speech Recognition models from Monolingual data sources
Viaarxiv icon

Fast Conformer with Linearly Scalable Attention for Efficient Speech Recognition

Add code
May 19, 2023
Figure 1 for Fast Conformer with Linearly Scalable Attention for Efficient Speech Recognition
Figure 2 for Fast Conformer with Linearly Scalable Attention for Efficient Speech Recognition
Figure 3 for Fast Conformer with Linearly Scalable Attention for Efficient Speech Recognition
Figure 4 for Fast Conformer with Linearly Scalable Attention for Efficient Speech Recognition
Viaarxiv icon