Picture for Yasaman Bahri

Yasaman Bahri

On the Emergence of Linear Analogies in Word Embeddings

Add code
May 24, 2025
Viaarxiv icon

Quantum Many-Body Physics Calculations with Large Language Models

Add code
Mar 05, 2024
Figure 1 for Quantum Many-Body Physics Calculations with Large Language Models
Figure 2 for Quantum Many-Body Physics Calculations with Large Language Models
Figure 3 for Quantum Many-Body Physics Calculations with Large Language Models
Figure 4 for Quantum Many-Body Physics Calculations with Large Language Models
Viaarxiv icon

Les Houches Lectures on Deep Learning at Large & Infinite Width

Add code
Sep 08, 2023
Viaarxiv icon

Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models

Add code
Jun 10, 2022
Viaarxiv icon

The Evolution of Out-of-Distribution Robustness Throughout Fine-Tuning

Add code
Jun 30, 2021
Figure 1 for The Evolution of Out-of-Distribution Robustness Throughout Fine-Tuning
Figure 2 for The Evolution of Out-of-Distribution Robustness Throughout Fine-Tuning
Figure 3 for The Evolution of Out-of-Distribution Robustness Throughout Fine-Tuning
Figure 4 for The Evolution of Out-of-Distribution Robustness Throughout Fine-Tuning
Viaarxiv icon

Explaining Neural Scaling Laws

Add code
Feb 12, 2021
Figure 1 for Explaining Neural Scaling Laws
Figure 2 for Explaining Neural Scaling Laws
Figure 3 for Explaining Neural Scaling Laws
Figure 4 for Explaining Neural Scaling Laws
Viaarxiv icon

Exact posterior distributions of wide Bayesian neural networks

Add code
Jun 18, 2020
Figure 1 for Exact posterior distributions of wide Bayesian neural networks
Figure 2 for Exact posterior distributions of wide Bayesian neural networks
Viaarxiv icon

Infinite attention: NNGP and NTK for deep attention networks

Add code
Jun 18, 2020
Figure 1 for Infinite attention: NNGP and NTK for deep attention networks
Figure 2 for Infinite attention: NNGP and NTK for deep attention networks
Figure 3 for Infinite attention: NNGP and NTK for deep attention networks
Figure 4 for Infinite attention: NNGP and NTK for deep attention networks
Viaarxiv icon

The large learning rate phase of deep learning: the catapult mechanism

Add code
Mar 04, 2020
Figure 1 for The large learning rate phase of deep learning: the catapult mechanism
Figure 2 for The large learning rate phase of deep learning: the catapult mechanism
Figure 3 for The large learning rate phase of deep learning: the catapult mechanism
Figure 4 for The large learning rate phase of deep learning: the catapult mechanism
Viaarxiv icon

Wide Neural Networks of Any Depth Evolve as Linear Models Under Gradient Descent

Add code
Feb 18, 2019
Figure 1 for Wide Neural Networks of Any Depth Evolve as Linear Models Under Gradient Descent
Figure 2 for Wide Neural Networks of Any Depth Evolve as Linear Models Under Gradient Descent
Figure 3 for Wide Neural Networks of Any Depth Evolve as Linear Models Under Gradient Descent
Figure 4 for Wide Neural Networks of Any Depth Evolve as Linear Models Under Gradient Descent
Viaarxiv icon