Picture for Seijin Kobayashi

Seijin Kobayashi

Learning Randomized Algorithms with Transformers

Add code
Aug 20, 2024
Figure 1 for Learning Randomized Algorithms with Transformers
Figure 2 for Learning Randomized Algorithms with Transformers
Figure 3 for Learning Randomized Algorithms with Transformers
Figure 4 for Learning Randomized Algorithms with Transformers
Viaarxiv icon

When can transformers compositionally generalize in-context?

Add code
Jul 17, 2024
Figure 1 for When can transformers compositionally generalize in-context?
Figure 2 for When can transformers compositionally generalize in-context?
Figure 3 for When can transformers compositionally generalize in-context?
Figure 4 for When can transformers compositionally generalize in-context?
Viaarxiv icon

Attention as a Hypernetwork

Add code
Jun 09, 2024
Viaarxiv icon

Discovering modular solutions that generalize compositionally

Add code
Dec 22, 2023
Figure 1 for Discovering modular solutions that generalize compositionally
Figure 2 for Discovering modular solutions that generalize compositionally
Figure 3 for Discovering modular solutions that generalize compositionally
Figure 4 for Discovering modular solutions that generalize compositionally
Viaarxiv icon

Uncovering mesa-optimization algorithms in Transformers

Add code
Sep 11, 2023
Figure 1 for Uncovering mesa-optimization algorithms in Transformers
Figure 2 for Uncovering mesa-optimization algorithms in Transformers
Figure 3 for Uncovering mesa-optimization algorithms in Transformers
Figure 4 for Uncovering mesa-optimization algorithms in Transformers
Viaarxiv icon

Gated recurrent neural networks discover attention

Add code
Sep 04, 2023
Figure 1 for Gated recurrent neural networks discover attention
Figure 2 for Gated recurrent neural networks discover attention
Figure 3 for Gated recurrent neural networks discover attention
Figure 4 for Gated recurrent neural networks discover attention
Viaarxiv icon

Would I have gotten that reward? Long-term credit assignment by counterfactual contribution analysis

Add code
Jun 29, 2023
Figure 1 for Would I have gotten that reward? Long-term credit assignment by counterfactual contribution analysis
Figure 2 for Would I have gotten that reward? Long-term credit assignment by counterfactual contribution analysis
Figure 3 for Would I have gotten that reward? Long-term credit assignment by counterfactual contribution analysis
Figure 4 for Would I have gotten that reward? Long-term credit assignment by counterfactual contribution analysis
Viaarxiv icon

Disentangling the Predictive Variance of Deep Ensembles through the Neural Tangent Kernel

Add code
Oct 18, 2022
Figure 1 for Disentangling the Predictive Variance of Deep Ensembles through the Neural Tangent Kernel
Figure 2 for Disentangling the Predictive Variance of Deep Ensembles through the Neural Tangent Kernel
Figure 3 for Disentangling the Predictive Variance of Deep Ensembles through the Neural Tangent Kernel
Figure 4 for Disentangling the Predictive Variance of Deep Ensembles through the Neural Tangent Kernel
Viaarxiv icon

Meta-Learning via Classifier(-free) Guidance

Add code
Oct 17, 2022
Figure 1 for Meta-Learning via Classifier(-free) Guidance
Figure 2 for Meta-Learning via Classifier(-free) Guidance
Figure 3 for Meta-Learning via Classifier(-free) Guidance
Figure 4 for Meta-Learning via Classifier(-free) Guidance
Viaarxiv icon

The least-control principle for learning at equilibrium

Add code
Jul 04, 2022
Figure 1 for The least-control principle for learning at equilibrium
Figure 2 for The least-control principle for learning at equilibrium
Figure 3 for The least-control principle for learning at equilibrium
Figure 4 for The least-control principle for learning at equilibrium
Viaarxiv icon