Picture for Fei Jia

Fei Jia

HUAWEI

NVIDIA Nemotron 3: Efficient and Open Intelligence

Add code
Dec 24, 2025
Viaarxiv icon

Nemotron 3 Nano: Open, Efficient Mixture-of-Experts Hybrid Mamba-Transformer Model for Agentic Reasoning

Add code
Dec 23, 2025
Viaarxiv icon

Llama-Nemotron: Efficient Reasoning Models

Add code
May 02, 2025
Figure 1 for Llama-Nemotron: Efficient Reasoning Models
Figure 2 for Llama-Nemotron: Efficient Reasoning Models
Figure 3 for Llama-Nemotron: Efficient Reasoning Models
Figure 4 for Llama-Nemotron: Efficient Reasoning Models
Viaarxiv icon

SWAN-GPT: An Efficient and Scalable Approach for Long-Context Language Modeling

Add code
Apr 11, 2025
Viaarxiv icon

Nemotron-H: A Family of Accurate and Efficient Hybrid Mamba-Transformer Models

Add code
Apr 10, 2025
Figure 1 for Nemotron-H: A Family of Accurate and Efficient Hybrid Mamba-Transformer Models
Figure 2 for Nemotron-H: A Family of Accurate and Efficient Hybrid Mamba-Transformer Models
Figure 3 for Nemotron-H: A Family of Accurate and Efficient Hybrid Mamba-Transformer Models
Figure 4 for Nemotron-H: A Family of Accurate and Efficient Hybrid Mamba-Transformer Models
Viaarxiv icon

Automatically Planning Optimal Parallel Strategy for Large Language Models

Add code
Dec 31, 2024
Figure 1 for Automatically Planning Optimal Parallel Strategy for Large Language Models
Figure 2 for Automatically Planning Optimal Parallel Strategy for Large Language Models
Figure 3 for Automatically Planning Optimal Parallel Strategy for Large Language Models
Figure 4 for Automatically Planning Optimal Parallel Strategy for Large Language Models
Viaarxiv icon

Star Attention: Efficient LLM Inference over Long Sequences

Add code
Nov 26, 2024
Figure 1 for Star Attention: Efficient LLM Inference over Long Sequences
Figure 2 for Star Attention: Efficient LLM Inference over Long Sequences
Figure 3 for Star Attention: Efficient LLM Inference over Long Sequences
Figure 4 for Star Attention: Efficient LLM Inference over Long Sequences
Viaarxiv icon

Romanization Encoding For Multilingual ASR

Add code
Jul 05, 2024
Figure 1 for Romanization Encoding For Multilingual ASR
Figure 2 for Romanization Encoding For Multilingual ASR
Figure 3 for Romanization Encoding For Multilingual ASR
Figure 4 for Romanization Encoding For Multilingual ASR
Viaarxiv icon

RULER: What's the Real Context Size of Your Long-Context Language Models?

Add code
Apr 11, 2024
Figure 1 for RULER: What's the Real Context Size of Your Long-Context Language Models?
Figure 2 for RULER: What's the Real Context Size of Your Long-Context Language Models?
Figure 3 for RULER: What's the Real Context Size of Your Long-Context Language Models?
Figure 4 for RULER: What's the Real Context Size of Your Long-Context Language Models?
Viaarxiv icon

Transducers with Pronunciation-aware Embeddings for Automatic Speech Recognition

Add code
Apr 04, 2024
Figure 1 for Transducers with Pronunciation-aware Embeddings for Automatic Speech Recognition
Figure 2 for Transducers with Pronunciation-aware Embeddings for Automatic Speech Recognition
Figure 3 for Transducers with Pronunciation-aware Embeddings for Automatic Speech Recognition
Figure 4 for Transducers with Pronunciation-aware Embeddings for Automatic Speech Recognition
Viaarxiv icon