Alert button
Picture for Szymon Tworkowski

Szymon Tworkowski

Alert button

Analysing The Impact of Sequence Composition on Language Model Pre-Training

Add code
Bookmark button
Alert button
Feb 21, 2024
Yu Zhao, Yuanbin Qu, Konrad Staniszewski, Szymon Tworkowski, Wei Liu, Piotr Miłoś, Yuxiang Wu, Pasquale Minervini

Viaarxiv icon

Structured Packing in LLM Training Improves Long Context Utilization

Add code
Bookmark button
Alert button
Jan 02, 2024
Konrad Staniszewski, Szymon Tworkowski, Sebastian Jaszczur, Henryk Michalewski, Łukasz Kuciński, Piotr Miłoś

Viaarxiv icon

Explaining Competitive-Level Programming Solutions using LLMs

Add code
Bookmark button
Alert button
Jul 11, 2023
Jierui Li, Szymon Tworkowski, Yingying Wu, Raymond Mooney

Figure 1 for Explaining Competitive-Level Programming Solutions using LLMs
Figure 2 for Explaining Competitive-Level Programming Solutions using LLMs
Figure 3 for Explaining Competitive-Level Programming Solutions using LLMs
Figure 4 for Explaining Competitive-Level Programming Solutions using LLMs
Viaarxiv icon

Focused Transformer: Contrastive Training for Context Scaling

Add code
Bookmark button
Alert button
Jul 06, 2023
Szymon Tworkowski, Konrad Staniszewski, Mikołaj Pacek, Yuhuai Wu, Henryk Michalewski, Piotr Miłoś

Figure 1 for Focused Transformer: Contrastive Training for Context Scaling
Figure 2 for Focused Transformer: Contrastive Training for Context Scaling
Figure 3 for Focused Transformer: Contrastive Training for Context Scaling
Figure 4 for Focused Transformer: Contrastive Training for Context Scaling
Viaarxiv icon

Magnushammer: A Transformer-based Approach to Premise Selection

Add code
Bookmark button
Alert button
Mar 08, 2023
Maciej Mikuła, Szymon Antoniak, Szymon Tworkowski, Albert Qiaochu Jiang, Jin Peng Zhou, Christian Szegedy, Łukasz Kuciński, Piotr Miłoś, Yuhuai Wu

Figure 1 for Magnushammer: A Transformer-based Approach to Premise Selection
Figure 2 for Magnushammer: A Transformer-based Approach to Premise Selection
Figure 3 for Magnushammer: A Transformer-based Approach to Premise Selection
Figure 4 for Magnushammer: A Transformer-based Approach to Premise Selection
Viaarxiv icon

Thor: Wielding Hammers to Integrate Language Models and Automated Theorem Provers

Add code
Bookmark button
Alert button
May 22, 2022
Albert Q. Jiang, Wenda Li, Szymon Tworkowski, Konrad Czechowski, Tomasz Odrzygóźdź, Piotr Miłoś, Yuhuai Wu, Mateja Jamnik

Figure 1 for Thor: Wielding Hammers to Integrate Language Models and Automated Theorem Provers
Figure 2 for Thor: Wielding Hammers to Integrate Language Models and Automated Theorem Provers
Figure 3 for Thor: Wielding Hammers to Integrate Language Models and Automated Theorem Provers
Figure 4 for Thor: Wielding Hammers to Integrate Language Models and Automated Theorem Provers
Viaarxiv icon

Hierarchical Transformers Are More Efficient Language Models

Add code
Bookmark button
Alert button
Oct 26, 2021
Piotr Nawrot, Szymon Tworkowski, Michał Tyrolski, Łukasz Kaiser, Yuhuai Wu, Christian Szegedy, Henryk Michalewski

Figure 1 for Hierarchical Transformers Are More Efficient Language Models
Figure 2 for Hierarchical Transformers Are More Efficient Language Models
Figure 3 for Hierarchical Transformers Are More Efficient Language Models
Figure 4 for Hierarchical Transformers Are More Efficient Language Models
Viaarxiv icon