Alert button
Picture for Mats L. Richter

Mats L. Richter

Alert button

Simple and Scalable Strategies to Continually Pre-train Large Language Models

Add code
Bookmark button
Alert button
Mar 26, 2024
Adam Ibrahim, Benjamin Thérien, Kshitij Gupta, Mats L. Richter, Quentin Anthony, Timothée Lesort, Eugene Belilovsky, Irina Rish

Figure 1 for Simple and Scalable Strategies to Continually Pre-train Large Language Models
Figure 2 for Simple and Scalable Strategies to Continually Pre-train Large Language Models
Figure 3 for Simple and Scalable Strategies to Continually Pre-train Large Language Models
Figure 4 for Simple and Scalable Strategies to Continually Pre-train Large Language Models
Viaarxiv icon

Continual Pre-Training of Large Language Models: How to (re)warm your model?

Add code
Bookmark button
Alert button
Aug 08, 2023
Kshitij Gupta, Benjamin Thérien, Adam Ibrahim, Mats L. Richter, Quentin Anthony, Eugene Belilovsky, Irina Rish, Timothée Lesort

Figure 1 for Continual Pre-Training of Large Language Models: How to (re)warm your model?
Figure 2 for Continual Pre-Training of Large Language Models: How to (re)warm your model?
Figure 3 for Continual Pre-Training of Large Language Models: How to (re)warm your model?
Figure 4 for Continual Pre-Training of Large Language Models: How to (re)warm your model?
Viaarxiv icon

Receptive Field Refinement for Convolutional Neural Networks Reliably Improves Predictive Performance

Add code
Bookmark button
Alert button
Nov 26, 2022
Mats L. Richter, Christopher Pal

Figure 1 for Receptive Field Refinement for Convolutional Neural Networks Reliably Improves Predictive Performance
Figure 2 for Receptive Field Refinement for Convolutional Neural Networks Reliably Improves Predictive Performance
Figure 3 for Receptive Field Refinement for Convolutional Neural Networks Reliably Improves Predictive Performance
Figure 4 for Receptive Field Refinement for Convolutional Neural Networks Reliably Improves Predictive Performance
Viaarxiv icon

Should You Go Deeper? Optimizing Convolutional Neural Network Architectures without Training by Receptive Field Analysis

Add code
Bookmark button
Alert button
Jun 23, 2021
Mats L. Richter, Julius Schöning, Ulf Krumnack

Figure 1 for Should You Go Deeper? Optimizing Convolutional Neural Network Architectures without Training by Receptive Field Analysis
Figure 2 for Should You Go Deeper? Optimizing Convolutional Neural Network Architectures without Training by Receptive Field Analysis
Figure 3 for Should You Go Deeper? Optimizing Convolutional Neural Network Architectures without Training by Receptive Field Analysis
Figure 4 for Should You Go Deeper? Optimizing Convolutional Neural Network Architectures without Training by Receptive Field Analysis
Viaarxiv icon

Exploring the Properties and Evolution of Neural Network Eigenspaces during Training

Add code
Bookmark button
Alert button
Jun 18, 2021
Mats L. Richter, Leila Malihi, Anne-Kathrin Patricia Windler, Ulf Krumnack

Figure 1 for Exploring the Properties and Evolution of Neural Network Eigenspaces during Training
Figure 2 for Exploring the Properties and Evolution of Neural Network Eigenspaces during Training
Figure 3 for Exploring the Properties and Evolution of Neural Network Eigenspaces during Training
Figure 4 for Exploring the Properties and Evolution of Neural Network Eigenspaces during Training
Viaarxiv icon

Size Matters

Add code
Bookmark button
Alert button
Feb 09, 2021
Mats L. Richter, Wolf Byttner, Ulf Krumnack, Ludwdig Schallner, Justin Shenk

Figure 1 for Size Matters
Figure 2 for Size Matters
Figure 3 for Size Matters
Figure 4 for Size Matters
Viaarxiv icon

Feature Space Saturation during Training

Add code
Bookmark button
Alert button
Jun 18, 2020
Justin Shenk, Mats L. Richter, Wolf Byttner, Anders Arpteg, Mikael Huss

Figure 1 for Feature Space Saturation during Training
Figure 2 for Feature Space Saturation during Training
Figure 3 for Feature Space Saturation during Training
Figure 4 for Feature Space Saturation during Training
Viaarxiv icon

Spectral Analysis of Latent Representations

Add code
Bookmark button
Alert button
Jul 19, 2019
Justin Shenk, Mats L. Richter, Anders Arpteg, Mikael Huss

Figure 1 for Spectral Analysis of Latent Representations
Figure 2 for Spectral Analysis of Latent Representations
Figure 3 for Spectral Analysis of Latent Representations
Figure 4 for Spectral Analysis of Latent Representations
Viaarxiv icon