Picture for Szymon Tworkowski

Szymon Tworkowski

Analysing The Impact of Sequence Composition on Language Model Pre-Training

Add code
Feb 21, 2024
Figure 1 for Analysing The Impact of Sequence Composition on Language Model Pre-Training
Figure 2 for Analysing The Impact of Sequence Composition on Language Model Pre-Training
Figure 3 for Analysing The Impact of Sequence Composition on Language Model Pre-Training
Figure 4 for Analysing The Impact of Sequence Composition on Language Model Pre-Training
Viaarxiv icon

Structured Packing in LLM Training Improves Long Context Utilization

Add code
Jan 02, 2024
Figure 1 for Structured Packing in LLM Training Improves Long Context Utilization
Figure 2 for Structured Packing in LLM Training Improves Long Context Utilization
Figure 3 for Structured Packing in LLM Training Improves Long Context Utilization
Figure 4 for Structured Packing in LLM Training Improves Long Context Utilization
Viaarxiv icon

Explaining Competitive-Level Programming Solutions using LLMs

Add code
Jul 11, 2023
Figure 1 for Explaining Competitive-Level Programming Solutions using LLMs
Figure 2 for Explaining Competitive-Level Programming Solutions using LLMs
Figure 3 for Explaining Competitive-Level Programming Solutions using LLMs
Figure 4 for Explaining Competitive-Level Programming Solutions using LLMs
Viaarxiv icon

Focused Transformer: Contrastive Training for Context Scaling

Add code
Jul 06, 2023
Figure 1 for Focused Transformer: Contrastive Training for Context Scaling
Figure 2 for Focused Transformer: Contrastive Training for Context Scaling
Figure 3 for Focused Transformer: Contrastive Training for Context Scaling
Figure 4 for Focused Transformer: Contrastive Training for Context Scaling
Viaarxiv icon

Magnushammer: A Transformer-based Approach to Premise Selection

Add code
Mar 08, 2023
Figure 1 for Magnushammer: A Transformer-based Approach to Premise Selection
Figure 2 for Magnushammer: A Transformer-based Approach to Premise Selection
Figure 3 for Magnushammer: A Transformer-based Approach to Premise Selection
Figure 4 for Magnushammer: A Transformer-based Approach to Premise Selection
Viaarxiv icon

Thor: Wielding Hammers to Integrate Language Models and Automated Theorem Provers

Add code
May 22, 2022
Figure 1 for Thor: Wielding Hammers to Integrate Language Models and Automated Theorem Provers
Figure 2 for Thor: Wielding Hammers to Integrate Language Models and Automated Theorem Provers
Figure 3 for Thor: Wielding Hammers to Integrate Language Models and Automated Theorem Provers
Figure 4 for Thor: Wielding Hammers to Integrate Language Models and Automated Theorem Provers
Viaarxiv icon

Hierarchical Transformers Are More Efficient Language Models

Add code
Oct 26, 2021
Figure 1 for Hierarchical Transformers Are More Efficient Language Models
Figure 2 for Hierarchical Transformers Are More Efficient Language Models
Figure 3 for Hierarchical Transformers Are More Efficient Language Models
Figure 4 for Hierarchical Transformers Are More Efficient Language Models
Viaarxiv icon