Alert button
Picture for Sandeep Subramanian

Sandeep Subramanian

Alert button

Nemotron-4 15B Technical Report

Add code
Bookmark button
Alert button
Feb 27, 2024
Jupinder Parmar, Shrimai Prabhumoye, Joseph Jennings, Mostofa Patwary, Sandeep Subramanian, Dan Su, Chen Zhu, Deepak Narayanan, Aastha Jhunjhunwala, Ayush Dattagupta, Vibhu Jawa, Jiwei Liu, Ameya Mahabaleshwarkar, Osvald Nitski, Annika Brundyn, James Maki, Miguel Martinez, Jiaxuan You, John Kamalu, Patrick LeGresley, Denys Fridman, Jared Casper, Ashwath Aithal, Oleksii Kuchaiev, Mohammad Shoeybi, Jonathan Cohen, Bryan Catanzaro

Viaarxiv icon

Mixtral of Experts

Add code
Bookmark button
Alert button
Jan 08, 2024
Albert Q. Jiang, Alexandre Sablayrolles, Antoine Roux, Arthur Mensch, Blanche Savary, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Emma Bou Hanna, Florian Bressand, Gianna Lengyel, Guillaume Bour, Guillaume Lample, Lélio Renard Lavaud, Lucile Saulnier, Marie-Anne Lachaux, Pierre Stock, Sandeep Subramanian, Sophia Yang, Szymon Antoniak, Teven Le Scao, Théophile Gervet, Thibaut Lavril, Thomas Wang, Timothée Lacroix, William El Sayed

Viaarxiv icon

Retrieval meets Long Context Large Language Models

Add code
Bookmark button
Alert button
Oct 04, 2023
Peng Xu, Wei Ping, Xianchao Wu, Lawrence McAfee, Chen Zhu, Zihan Liu, Sandeep Subramanian, Evelina Bakhturina, Mohammad Shoeybi, Bryan Catanzaro

Figure 1 for Retrieval meets Long Context Large Language Models
Figure 2 for Retrieval meets Long Context Large Language Models
Figure 3 for Retrieval meets Long Context Large Language Models
Figure 4 for Retrieval meets Long Context Large Language Models
Viaarxiv icon

Finding the Right Recipe for Low Resource Domain Adaptation in Neural Machine Translation

Add code
Bookmark button
Alert button
Jun 02, 2022
Virginia Adams, Sandeep Subramanian, Mike Chrzanowski, Oleksii Hrinchuk, Oleksii Kuchaiev

Figure 1 for Finding the Right Recipe for Low Resource Domain Adaptation in Neural Machine Translation
Figure 2 for Finding the Right Recipe for Low Resource Domain Adaptation in Neural Machine Translation
Figure 3 for Finding the Right Recipe for Low Resource Domain Adaptation in Neural Machine Translation
Figure 4 for Finding the Right Recipe for Low Resource Domain Adaptation in Neural Machine Translation
Viaarxiv icon

NVIDIA NeMo Neural Machine Translation Systems for English-German and English-Russian News and Biomedical Tasks at WMT21

Add code
Bookmark button
Alert button
Nov 16, 2021
Sandeep Subramanian, Oleksii Hrinchuk, Virginia Adams, Oleksii Kuchaiev

Figure 1 for NVIDIA NeMo Neural Machine Translation Systems for English-German and English-Russian News and Biomedical Tasks at WMT21
Figure 2 for NVIDIA NeMo Neural Machine Translation Systems for English-German and English-Russian News and Biomedical Tasks at WMT21
Figure 3 for NVIDIA NeMo Neural Machine Translation Systems for English-German and English-Russian News and Biomedical Tasks at WMT21
Figure 4 for NVIDIA NeMo Neural Machine Translation Systems for English-German and English-Russian News and Biomedical Tasks at WMT21
Viaarxiv icon

Multi-scale Transformer Language Models

Add code
Bookmark button
Alert button
May 01, 2020
Sandeep Subramanian, Ronan Collobert, Marc'Aurelio Ranzato, Y-Lan Boureau

Figure 1 for Multi-scale Transformer Language Models
Figure 2 for Multi-scale Transformer Language Models
Figure 3 for Multi-scale Transformer Language Models
Figure 4 for Multi-scale Transformer Language Models
Viaarxiv icon

On Extractive and Abstractive Neural Document Summarization with Transformer Language Models

Add code
Bookmark button
Alert button
Sep 07, 2019
Sandeep Subramanian, Raymond Li, Jonathan Pilault, Christopher Pal

Figure 1 for On Extractive and Abstractive Neural Document Summarization with Transformer Language Models
Figure 2 for On Extractive and Abstractive Neural Document Summarization with Transformer Language Models
Figure 3 for On Extractive and Abstractive Neural Document Summarization with Transformer Language Models
Figure 4 for On Extractive and Abstractive Neural Document Summarization with Transformer Language Models
Viaarxiv icon

Do Neural Dialog Systems Use the Conversation History Effectively? An Empirical Study

Add code
Bookmark button
Alert button
Jun 04, 2019
Chinnadhurai Sankar, Sandeep Subramanian, Christopher Pal, Sarath Chandar, Yoshua Bengio

Figure 1 for Do Neural Dialog Systems Use the Conversation History Effectively? An Empirical Study
Figure 2 for Do Neural Dialog Systems Use the Conversation History Effectively? An Empirical Study
Figure 3 for Do Neural Dialog Systems Use the Conversation History Effectively? An Empirical Study
Viaarxiv icon

State-Reification Networks: Improving Generalization by Modeling the Distribution of Hidden Representations

Add code
Bookmark button
Alert button
May 26, 2019
Alex Lamb, Jonathan Binas, Anirudh Goyal, Sandeep Subramanian, Ioannis Mitliagkas, Denis Kazakov, Yoshua Bengio, Michael C. Mozer

Figure 1 for State-Reification Networks: Improving Generalization by Modeling the Distribution of Hidden Representations
Figure 2 for State-Reification Networks: Improving Generalization by Modeling the Distribution of Hidden Representations
Figure 3 for State-Reification Networks: Improving Generalization by Modeling the Distribution of Hidden Representations
Figure 4 for State-Reification Networks: Improving Generalization by Modeling the Distribution of Hidden Representations
Viaarxiv icon

Multiple-Attribute Text Style Transfer

Add code
Bookmark button
Alert button
Nov 01, 2018
Sandeep Subramanian, Guillaume Lample, Eric Michael Smith, Ludovic Denoyer, Marc'Aurelio Ranzato, Y-Lan Boureau

Figure 1 for Multiple-Attribute Text Style Transfer
Figure 2 for Multiple-Attribute Text Style Transfer
Figure 3 for Multiple-Attribute Text Style Transfer
Figure 4 for Multiple-Attribute Text Style Transfer
Viaarxiv icon