Picture for Matthias Bethge

Matthias Bethge

Adaptation Odyssey in LLMs: Why Does Additional Pretraining Sometimes Fail to Improve?

Add code
Oct 08, 2024
Figure 1 for Adaptation Odyssey in LLMs: Why Does Additional Pretraining Sometimes Fail to Improve?
Figure 2 for Adaptation Odyssey in LLMs: Why Does Additional Pretraining Sometimes Fail to Improve?
Figure 3 for Adaptation Odyssey in LLMs: Why Does Additional Pretraining Sometimes Fail to Improve?
Figure 4 for Adaptation Odyssey in LLMs: Why Does Additional Pretraining Sometimes Fail to Improve?
Viaarxiv icon

A Practitioner's Guide to Continual Multimodal Pretraining

Add code
Aug 26, 2024
Viaarxiv icon

Reflecting on the State of Rehearsal-free Continual Learning with Pretrained Models

Add code
Jun 13, 2024
Viaarxiv icon

Identifying latent state transition in non-linear dynamical systems

Add code
Jun 06, 2024
Viaarxiv icon

The Entropy Enigma: Success and Failure of Entropy Minimization

Add code
May 08, 2024
Viaarxiv icon

Wu's Method can Boost Symbolic AI to Rival Silver Medalists and AlphaGeometry to Outperform Gold Medalists at IMO Geometry

Add code
Apr 11, 2024
Figure 1 for Wu's Method can Boost Symbolic AI to Rival Silver Medalists and AlphaGeometry to Outperform Gold Medalists at IMO Geometry
Figure 2 for Wu's Method can Boost Symbolic AI to Rival Silver Medalists and AlphaGeometry to Outperform Gold Medalists at IMO Geometry
Figure 3 for Wu's Method can Boost Symbolic AI to Rival Silver Medalists and AlphaGeometry to Outperform Gold Medalists at IMO Geometry
Viaarxiv icon

No "Zero-Shot" Without Exponential Data: Pretraining Concept Frequency Determines Multimodal Model Performance

Add code
Apr 08, 2024
Figure 1 for No "Zero-Shot" Without Exponential Data: Pretraining Concept Frequency Determines Multimodal Model Performance
Figure 2 for No "Zero-Shot" Without Exponential Data: Pretraining Concept Frequency Determines Multimodal Model Performance
Figure 3 for No "Zero-Shot" Without Exponential Data: Pretraining Concept Frequency Determines Multimodal Model Performance
Figure 4 for No "Zero-Shot" Without Exponential Data: Pretraining Concept Frequency Determines Multimodal Model Performance
Viaarxiv icon

Lifelong Benchmarks: Efficient Model Evaluation in an Era of Rapid Progress

Add code
Feb 29, 2024
Figure 1 for Lifelong Benchmarks: Efficient Model Evaluation in an Era of Rapid Progress
Figure 2 for Lifelong Benchmarks: Efficient Model Evaluation in an Era of Rapid Progress
Figure 3 for Lifelong Benchmarks: Efficient Model Evaluation in an Era of Rapid Progress
Figure 4 for Lifelong Benchmarks: Efficient Model Evaluation in an Era of Rapid Progress
Viaarxiv icon

Investigating Continual Pretraining in Large Language Models: Insights and Implications

Add code
Feb 27, 2024
Figure 1 for Investigating Continual Pretraining in Large Language Models: Insights and Implications
Figure 2 for Investigating Continual Pretraining in Large Language Models: Insights and Implications
Figure 3 for Investigating Continual Pretraining in Large Language Models: Insights and Implications
Figure 4 for Investigating Continual Pretraining in Large Language Models: Insights and Implications
Viaarxiv icon

Disentangled Continual Learning: Separating Memory Edits from Model Updates

Add code
Dec 27, 2023
Figure 1 for Disentangled Continual Learning: Separating Memory Edits from Model Updates
Figure 2 for Disentangled Continual Learning: Separating Memory Edits from Model Updates
Figure 3 for Disentangled Continual Learning: Separating Memory Edits from Model Updates
Figure 4 for Disentangled Continual Learning: Separating Memory Edits from Model Updates
Viaarxiv icon