Picture for Ankush Garg

Ankush Garg

Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context

Add code
Mar 08, 2024
Viaarxiv icon

Gemini: A Family of Highly Capable Multimodal Models

Add code
Dec 19, 2023
Viaarxiv icon

Order Matters in the Presence of Dataset Imbalance for Multilingual Learning

Add code
Dec 11, 2023
Viaarxiv icon

The Devil is in the Errors: Leveraging Large Language Models for Fine-grained Machine Translation Evaluation

Add code
Aug 14, 2023
Figure 1 for The Devil is in the Errors: Leveraging Large Language Models for Fine-grained Machine Translation Evaluation
Figure 2 for The Devil is in the Errors: Leveraging Large Language Models for Fine-grained Machine Translation Evaluation
Figure 3 for The Devil is in the Errors: Leveraging Large Language Models for Fine-grained Machine Translation Evaluation
Figure 4 for The Devil is in the Errors: Leveraging Large Language Models for Fine-grained Machine Translation Evaluation
Viaarxiv icon

Benchmarking Neural Network Training Algorithms

Add code
Jun 12, 2023
Figure 1 for Benchmarking Neural Network Training Algorithms
Figure 2 for Benchmarking Neural Network Training Algorithms
Figure 3 for Benchmarking Neural Network Training Algorithms
Figure 4 for Benchmarking Neural Network Training Algorithms
Viaarxiv icon

Binarized Neural Machine Translation

Add code
Feb 09, 2023
Figure 1 for Binarized Neural Machine Translation
Figure 2 for Binarized Neural Machine Translation
Figure 3 for Binarized Neural Machine Translation
Figure 4 for Binarized Neural Machine Translation
Viaarxiv icon

Do Current Multi-Task Optimization Methods in Deep Learning Even Help?

Add code
Sep 23, 2022
Figure 1 for Do Current Multi-Task Optimization Methods in Deep Learning Even Help?
Figure 2 for Do Current Multi-Task Optimization Methods in Deep Learning Even Help?
Figure 3 for Do Current Multi-Task Optimization Methods in Deep Learning Even Help?
Figure 4 for Do Current Multi-Task Optimization Methods in Deep Learning Even Help?
Viaarxiv icon

Data Scaling Laws in NMT: The Effect of Noise and Architecture

Add code
Feb 04, 2022
Figure 1 for Data Scaling Laws in NMT: The Effect of Noise and Architecture
Figure 2 for Data Scaling Laws in NMT: The Effect of Noise and Architecture
Figure 3 for Data Scaling Laws in NMT: The Effect of Noise and Architecture
Figure 4 for Data Scaling Laws in NMT: The Effect of Noise and Architecture
Viaarxiv icon

A Loss Curvature Perspective on Training Instability in Deep Learning

Add code
Oct 08, 2021
Figure 1 for A Loss Curvature Perspective on Training Instability in Deep Learning
Figure 2 for A Loss Curvature Perspective on Training Instability in Deep Learning
Figure 3 for A Loss Curvature Perspective on Training Instability in Deep Learning
Figure 4 for A Loss Curvature Perspective on Training Instability in Deep Learning
Viaarxiv icon

Efficient and Private Federated Learning with Partially Trainable Networks

Add code
Oct 06, 2021
Figure 1 for Efficient and Private Federated Learning with Partially Trainable Networks
Figure 2 for Efficient and Private Federated Learning with Partially Trainable Networks
Figure 3 for Efficient and Private Federated Learning with Partially Trainable Networks
Figure 4 for Efficient and Private Federated Learning with Partially Trainable Networks
Viaarxiv icon