Picture for Adhiguna Kuncoro

Adhiguna Kuncoro

DiPaCo: Distributed Path Composition

Add code
Mar 15, 2024
Figure 1 for DiPaCo: Distributed Path Composition
Figure 2 for DiPaCo: Distributed Path Composition
Figure 3 for DiPaCo: Distributed Path Composition
Figure 4 for DiPaCo: Distributed Path Composition
Viaarxiv icon

DiLoCo: Distributed Low-Communication Training of Language Models

Add code
Nov 14, 2023
Figure 1 for DiLoCo: Distributed Low-Communication Training of Language Models
Figure 2 for DiLoCo: Distributed Low-Communication Training of Language Models
Figure 3 for DiLoCo: Distributed Low-Communication Training of Language Models
Figure 4 for DiLoCo: Distributed Low-Communication Training of Language Models
Viaarxiv icon

On "Scientific Debt" in NLP: A Case for More Rigour in Language Model Pre-Training Research

Add code
Jun 05, 2023
Figure 1 for On "Scientific Debt" in NLP: A Case for More Rigour in Language Model Pre-Training Research
Figure 2 for On "Scientific Debt" in NLP: A Case for More Rigour in Language Model Pre-Training Research
Figure 3 for On "Scientific Debt" in NLP: A Case for More Rigour in Language Model Pre-Training Research
Figure 4 for On "Scientific Debt" in NLP: A Case for More Rigour in Language Model Pre-Training Research
Viaarxiv icon

A Natural Bias for Language Generation Models

Add code
Dec 19, 2022
Figure 1 for A Natural Bias for Language Generation Models
Figure 2 for A Natural Bias for Language Generation Models
Figure 3 for A Natural Bias for Language Generation Models
Figure 4 for A Natural Bias for Language Generation Models
Viaarxiv icon

Transformer Grammars: Augmenting Transformer Language Models with Syntactic Inductive Biases at Scale

Add code
Mar 01, 2022
Figure 1 for Transformer Grammars: Augmenting Transformer Language Models with Syntactic Inductive Biases at Scale
Figure 2 for Transformer Grammars: Augmenting Transformer Language Models with Syntactic Inductive Biases at Scale
Figure 3 for Transformer Grammars: Augmenting Transformer Language Models with Syntactic Inductive Biases at Scale
Figure 4 for Transformer Grammars: Augmenting Transformer Language Models with Syntactic Inductive Biases at Scale
Viaarxiv icon

Scaling Language Models: Methods, Analysis & Insights from Training Gopher

Add code
Dec 08, 2021
Figure 1 for Scaling Language Models: Methods, Analysis & Insights from Training Gopher
Figure 2 for Scaling Language Models: Methods, Analysis & Insights from Training Gopher
Figure 3 for Scaling Language Models: Methods, Analysis & Insights from Training Gopher
Figure 4 for Scaling Language Models: Methods, Analysis & Insights from Training Gopher
Viaarxiv icon

IndoNLG: Benchmark and Resources for Evaluating Indonesian Natural Language Generation

Add code
Apr 16, 2021
Figure 1 for IndoNLG: Benchmark and Resources for Evaluating Indonesian Natural Language Generation
Figure 2 for IndoNLG: Benchmark and Resources for Evaluating Indonesian Natural Language Generation
Figure 3 for IndoNLG: Benchmark and Resources for Evaluating Indonesian Natural Language Generation
Figure 4 for IndoNLG: Benchmark and Resources for Evaluating Indonesian Natural Language Generation
Viaarxiv icon

Pitfalls of Static Language Modelling

Add code
Feb 03, 2021
Figure 1 for Pitfalls of Static Language Modelling
Figure 2 for Pitfalls of Static Language Modelling
Figure 3 for Pitfalls of Static Language Modelling
Figure 4 for Pitfalls of Static Language Modelling
Viaarxiv icon

Syntactic Structure Distillation Pretraining For Bidirectional Encoders

Add code
May 27, 2020
Figure 1 for Syntactic Structure Distillation Pretraining For Bidirectional Encoders
Figure 2 for Syntactic Structure Distillation Pretraining For Bidirectional Encoders
Figure 3 for Syntactic Structure Distillation Pretraining For Bidirectional Encoders
Figure 4 for Syntactic Structure Distillation Pretraining For Bidirectional Encoders
Viaarxiv icon

Scalable Syntax-Aware Language Models Using Knowledge Distillation

Add code
Jun 14, 2019
Figure 1 for Scalable Syntax-Aware Language Models Using Knowledge Distillation
Figure 2 for Scalable Syntax-Aware Language Models Using Knowledge Distillation
Figure 3 for Scalable Syntax-Aware Language Models Using Knowledge Distillation
Figure 4 for Scalable Syntax-Aware Language Models Using Knowledge Distillation
Viaarxiv icon