Picture for Hidenori Tanaka

Hidenori Tanaka

Emergence of Hidden Capabilities: Exploring Learning Dynamics in Concept Space

Add code
Jun 27, 2024
Viaarxiv icon

Towards an Understanding of Stepwise Inference in Transformers: A Synthetic Graph Navigation Model

Add code
Feb 12, 2024
Viaarxiv icon

How Capable Can a Transformer Become? A Study on Synthetic, Interpretable Tasks

Add code
Nov 21, 2023
Viaarxiv icon

Mechanistically analyzing the effects of fine-tuning on procedurally defined tasks

Add code
Nov 21, 2023
Viaarxiv icon

In-Context Learning Dynamics with Random Binary Sequences

Add code
Oct 26, 2023
Figure 1 for In-Context Learning Dynamics with Random Binary Sequences
Figure 2 for In-Context Learning Dynamics with Random Binary Sequences
Figure 3 for In-Context Learning Dynamics with Random Binary Sequences
Figure 4 for In-Context Learning Dynamics with Random Binary Sequences
Viaarxiv icon

Compositional Abilities Emerge Multiplicatively: Exploring Diffusion Models on a Synthetic Task

Add code
Oct 13, 2023
Figure 1 for Compositional Abilities Emerge Multiplicatively: Exploring Diffusion Models on a Synthetic Task
Figure 2 for Compositional Abilities Emerge Multiplicatively: Exploring Diffusion Models on a Synthetic Task
Figure 3 for Compositional Abilities Emerge Multiplicatively: Exploring Diffusion Models on a Synthetic Task
Figure 4 for Compositional Abilities Emerge Multiplicatively: Exploring Diffusion Models on a Synthetic Task
Viaarxiv icon

Mechanistic Mode Connectivity

Add code
Nov 15, 2022
Figure 1 for Mechanistic Mode Connectivity
Figure 2 for Mechanistic Mode Connectivity
Figure 3 for Mechanistic Mode Connectivity
Figure 4 for Mechanistic Mode Connectivity
Viaarxiv icon

What shapes the loss landscape of self-supervised learning?

Add code
Oct 02, 2022
Figure 1 for What shapes the loss landscape of self-supervised learning?
Figure 2 for What shapes the loss landscape of self-supervised learning?
Figure 3 for What shapes the loss landscape of self-supervised learning?
Figure 4 for What shapes the loss landscape of self-supervised learning?
Viaarxiv icon

Rethinking the limiting dynamics of SGD: modified loss, phase space oscillations, and anomalous diffusion

Add code
Jul 19, 2021
Figure 1 for Rethinking the limiting dynamics of SGD: modified loss, phase space oscillations, and anomalous diffusion
Figure 2 for Rethinking the limiting dynamics of SGD: modified loss, phase space oscillations, and anomalous diffusion
Figure 3 for Rethinking the limiting dynamics of SGD: modified loss, phase space oscillations, and anomalous diffusion
Figure 4 for Rethinking the limiting dynamics of SGD: modified loss, phase space oscillations, and anomalous diffusion
Viaarxiv icon

Beyond BatchNorm: Towards a General Understanding of Normalization in Deep Learning

Add code
Jul 08, 2021
Figure 1 for Beyond BatchNorm: Towards a General Understanding of Normalization in Deep Learning
Figure 2 for Beyond BatchNorm: Towards a General Understanding of Normalization in Deep Learning
Figure 3 for Beyond BatchNorm: Towards a General Understanding of Normalization in Deep Learning
Figure 4 for Beyond BatchNorm: Towards a General Understanding of Normalization in Deep Learning
Viaarxiv icon