Alert button
Picture for Piotr Miłoś

Piotr Miłoś

Alert button

tsGT: Stochastic Time Series Modeling With Transformer

Add code
Bookmark button
Alert button
Mar 15, 2024
Łukasz Kuciński, Witold Drzewakowski, Mateusz Olko, Piotr Kozakowski, Łukasz Maziarka, Marta Emilia Nowakowska, Łukasz Kaiser, Piotr Miłoś

Figure 1 for tsGT: Stochastic Time Series Modeling With Transformer
Figure 2 for tsGT: Stochastic Time Series Modeling With Transformer
Figure 3 for tsGT: Stochastic Time Series Modeling With Transformer
Figure 4 for tsGT: Stochastic Time Series Modeling With Transformer
Viaarxiv icon

Overestimation, Overfitting, and Plasticity in Actor-Critic: the Bitter Lesson of Reinforcement Learning

Add code
Bookmark button
Alert button
Mar 01, 2024
Michal Nauman, Michał Bortkiewicz, Mateusz Ostaszewski, Piotr Miłoś, Tomasz Trzciński, Marek Cygan

Figure 1 for Overestimation, Overfitting, and Plasticity in Actor-Critic: the Bitter Lesson of Reinforcement Learning
Figure 2 for Overestimation, Overfitting, and Plasticity in Actor-Critic: the Bitter Lesson of Reinforcement Learning
Figure 3 for Overestimation, Overfitting, and Plasticity in Actor-Critic: the Bitter Lesson of Reinforcement Learning
Figure 4 for Overestimation, Overfitting, and Plasticity in Actor-Critic: the Bitter Lesson of Reinforcement Learning
Viaarxiv icon

Analysing The Impact of Sequence Composition on Language Model Pre-Training

Add code
Bookmark button
Alert button
Feb 21, 2024
Yu Zhao, Yuanbin Qu, Konrad Staniszewski, Szymon Tworkowski, Wei Liu, Piotr Miłoś, Yuxiang Wu, Pasquale Minervini

Viaarxiv icon

Fine-tuning Reinforcement Learning Models is Secretly a Forgetting Mitigation Problem

Add code
Bookmark button
Alert button
Feb 05, 2024
Maciej Wołczyk, Bartłomiej Cupiał, Mateusz Ostaszewski, Michał Bortkiewicz, Michał Zając, Razvan Pascanu, Łukasz Kuciński, Piotr Miłoś

Viaarxiv icon

Structured Packing in LLM Training Improves Long Context Utilization

Add code
Bookmark button
Alert button
Jan 02, 2024
Konrad Staniszewski, Szymon Tworkowski, Sebastian Jaszczur, Henryk Michalewski, Łukasz Kuciński, Piotr Miłoś

Viaarxiv icon

Focused Transformer: Contrastive Training for Context Scaling

Add code
Bookmark button
Alert button
Jul 06, 2023
Szymon Tworkowski, Konrad Staniszewski, Mikołaj Pacek, Yuhuai Wu, Henryk Michalewski, Piotr Miłoś

Figure 1 for Focused Transformer: Contrastive Training for Context Scaling
Figure 2 for Focused Transformer: Contrastive Training for Context Scaling
Figure 3 for Focused Transformer: Contrastive Training for Context Scaling
Figure 4 for Focused Transformer: Contrastive Training for Context Scaling
Viaarxiv icon

The Tunnel Effect: Building Data Representations in Deep Neural Networks

Add code
Bookmark button
Alert button
May 31, 2023
Wojciech Masarczyk, Mateusz Ostaszewski, Ehsan Imani, Razvan Pascanu, Piotr Miłoś, Tomasz Trzciński

Figure 1 for The Tunnel Effect: Building Data Representations in Deep Neural Networks
Figure 2 for The Tunnel Effect: Building Data Representations in Deep Neural Networks
Figure 3 for The Tunnel Effect: Building Data Representations in Deep Neural Networks
Figure 4 for The Tunnel Effect: Building Data Representations in Deep Neural Networks
Viaarxiv icon

Exploring Continual Learning of Diffusion Models

Add code
Bookmark button
Alert button
Mar 27, 2023
Michał Zając, Kamil Deja, Anna Kuzina, Jakub M. Tomczak, Tomasz Trzciński, Florian Shkurti, Piotr Miłoś

Figure 1 for Exploring Continual Learning of Diffusion Models
Figure 2 for Exploring Continual Learning of Diffusion Models
Figure 3 for Exploring Continual Learning of Diffusion Models
Figure 4 for Exploring Continual Learning of Diffusion Models
Viaarxiv icon

Magnushammer: A Transformer-based Approach to Premise Selection

Add code
Bookmark button
Alert button
Mar 08, 2023
Maciej Mikuła, Szymon Antoniak, Szymon Tworkowski, Albert Qiaochu Jiang, Jin Peng Zhou, Christian Szegedy, Łukasz Kuciński, Piotr Miłoś, Yuhuai Wu

Figure 1 for Magnushammer: A Transformer-based Approach to Premise Selection
Figure 2 for Magnushammer: A Transformer-based Approach to Premise Selection
Figure 3 for Magnushammer: A Transformer-based Approach to Premise Selection
Figure 4 for Magnushammer: A Transformer-based Approach to Premise Selection
Viaarxiv icon

The Surprising Effectiveness of Latent World Models for Continual Reinforcement Learning

Add code
Bookmark button
Alert button
Nov 29, 2022
Samuel Kessler, Piotr Miłoś, Jack Parker-Holder, Stephen J. Roberts

Figure 1 for The Surprising Effectiveness of Latent World Models for Continual Reinforcement Learning
Figure 2 for The Surprising Effectiveness of Latent World Models for Continual Reinforcement Learning
Figure 3 for The Surprising Effectiveness of Latent World Models for Continual Reinforcement Learning
Figure 4 for The Surprising Effectiveness of Latent World Models for Continual Reinforcement Learning
Viaarxiv icon