Picture for Deepak Narayanan

Deepak Narayanan

Nemotron-4 340B Technical Report

Add code
Jun 17, 2024
Figure 1 for Nemotron-4 340B Technical Report
Figure 2 for Nemotron-4 340B Technical Report
Figure 3 for Nemotron-4 340B Technical Report
Figure 4 for Nemotron-4 340B Technical Report
Viaarxiv icon

An Empirical Study of Mamba-based Language Models

Add code
Jun 12, 2024
Viaarxiv icon

Nemotron-4 15B Technical Report

Add code
Feb 27, 2024
Figure 1 for Nemotron-4 15B Technical Report
Figure 2 for Nemotron-4 15B Technical Report
Figure 3 for Nemotron-4 15B Technical Report
Figure 4 for Nemotron-4 15B Technical Report
Viaarxiv icon

The Case for Co-Designing Model Architectures with Hardware

Add code
Jan 30, 2024
Viaarxiv icon

Packrat: Automatic Reconfiguration for Latency Minimization in CPU-based DNN Serving

Add code
Nov 30, 2023
Viaarxiv icon

Holistic Evaluation of Text-To-Image Models

Add code
Nov 07, 2023
Figure 1 for Holistic Evaluation of Text-To-Image Models
Figure 2 for Holistic Evaluation of Text-To-Image Models
Figure 3 for Holistic Evaluation of Text-To-Image Models
Figure 4 for Holistic Evaluation of Text-To-Image Models
Viaarxiv icon

MGit: A Model Versioning and Management System

Add code
Jul 14, 2023
Figure 1 for MGit: A Model Versioning and Management System
Figure 2 for MGit: A Model Versioning and Management System
Figure 3 for MGit: A Model Versioning and Management System
Figure 4 for MGit: A Model Versioning and Management System
Viaarxiv icon

Cheaply Evaluating Inference Efficiency Metrics for Autoregressive Transformer APIs

Add code
May 03, 2023
Figure 1 for Cheaply Evaluating Inference Efficiency Metrics for Autoregressive Transformer APIs
Figure 2 for Cheaply Evaluating Inference Efficiency Metrics for Autoregressive Transformer APIs
Figure 3 for Cheaply Evaluating Inference Efficiency Metrics for Autoregressive Transformer APIs
Figure 4 for Cheaply Evaluating Inference Efficiency Metrics for Autoregressive Transformer APIs
Viaarxiv icon

MegaBlocks: Efficient Sparse Training with Mixture-of-Experts

Add code
Nov 29, 2022
Figure 1 for MegaBlocks: Efficient Sparse Training with Mixture-of-Experts
Figure 2 for MegaBlocks: Efficient Sparse Training with Mixture-of-Experts
Figure 3 for MegaBlocks: Efficient Sparse Training with Mixture-of-Experts
Figure 4 for MegaBlocks: Efficient Sparse Training with Mixture-of-Experts
Viaarxiv icon

Holistic Evaluation of Language Models

Add code
Nov 16, 2022
Figure 1 for Holistic Evaluation of Language Models
Figure 2 for Holistic Evaluation of Language Models
Figure 3 for Holistic Evaluation of Language Models
Figure 4 for Holistic Evaluation of Language Models
Viaarxiv icon