Picture for Adityanarayanan Radhakrishnan

Adityanarayanan Radhakrishnan

Context-Scaling versus Task-Scaling in In-Context Learning

Add code
Oct 16, 2024
Viaarxiv icon

Emergence in non-neural models: grokking modular arithmetic via average gradient outer product

Add code
Jul 29, 2024
Viaarxiv icon

Linear Recursive Feature Machines provably recover low-rank matrices

Add code
Jan 09, 2024
Viaarxiv icon

Mechanism of feature learning in convolutional neural networks

Add code
Sep 01, 2023
Viaarxiv icon

Catapults in SGD: spikes in the training loss and their impact on generalization through feature learning

Add code
Jun 07, 2023
Viaarxiv icon

Feature learning in neural networks and kernel machines that recursively learn features

Add code
Dec 28, 2022
Viaarxiv icon

Transfer Learning with Kernel Methods

Add code
Nov 01, 2022
Viaarxiv icon

Quadratic models for understanding neural network dynamics

Add code
May 24, 2022
Figure 1 for Quadratic models for understanding neural network dynamics
Figure 2 for Quadratic models for understanding neural network dynamics
Figure 3 for Quadratic models for understanding neural network dynamics
Figure 4 for Quadratic models for understanding neural network dynamics
Viaarxiv icon

Wide and Deep Neural Networks Achieve Optimality for Classification

Add code
Apr 29, 2022
Figure 1 for Wide and Deep Neural Networks Achieve Optimality for Classification
Figure 2 for Wide and Deep Neural Networks Achieve Optimality for Classification
Figure 3 for Wide and Deep Neural Networks Achieve Optimality for Classification
Viaarxiv icon

Local Quadratic Convergence of Stochastic Gradient Descent with Adaptive Step Size

Add code
Dec 30, 2021
Figure 1 for Local Quadratic Convergence of Stochastic Gradient Descent with Adaptive Step Size
Figure 2 for Local Quadratic Convergence of Stochastic Gradient Descent with Adaptive Step Size
Viaarxiv icon