Picture for Benjamin L. Edelman

Benjamin L. Edelman

Transcendence: Generative Models Can Outperform The Experts That Train Them

Add code
Jun 17, 2024
Figure 1 for Transcendence: Generative Models Can Outperform The Experts That Train Them
Figure 2 for Transcendence: Generative Models Can Outperform The Experts That Train Them
Figure 3 for Transcendence: Generative Models Can Outperform The Experts That Train Them
Figure 4 for Transcendence: Generative Models Can Outperform The Experts That Train Them
Viaarxiv icon

Foundational Challenges in Assuring Alignment and Safety of Large Language Models

Add code
Apr 15, 2024
Viaarxiv icon

The Evolution of Statistical Induction Heads: In-Context Learning Markov Chains

Add code
Feb 16, 2024
Viaarxiv icon

Distinguishing the Knowable from the Unknowable with Language Models

Add code
Feb 05, 2024
Viaarxiv icon

Watermarks in the Sand: Impossibility of Strong Watermarking for Generative Models

Add code
Nov 15, 2023
Viaarxiv icon

Feature emergence via margin maximization: case studies in algebraic tasks

Add code
Nov 13, 2023
Figure 1 for Feature emergence via margin maximization: case studies in algebraic tasks
Figure 2 for Feature emergence via margin maximization: case studies in algebraic tasks
Figure 3 for Feature emergence via margin maximization: case studies in algebraic tasks
Figure 4 for Feature emergence via margin maximization: case studies in algebraic tasks
Viaarxiv icon

Pareto Frontiers in Neural Feature Learning: Data, Compute, Width, and Luck

Add code
Sep 07, 2023
Figure 1 for Pareto Frontiers in Neural Feature Learning: Data, Compute, Width, and Luck
Figure 2 for Pareto Frontiers in Neural Feature Learning: Data, Compute, Width, and Luck
Figure 3 for Pareto Frontiers in Neural Feature Learning: Data, Compute, Width, and Luck
Figure 4 for Pareto Frontiers in Neural Feature Learning: Data, Compute, Width, and Luck
Viaarxiv icon

Hidden Progress in Deep Learning: SGD Learns Parities Near the Computational Limit

Add code
Jul 18, 2022
Figure 1 for Hidden Progress in Deep Learning: SGD Learns Parities Near the Computational Limit
Figure 2 for Hidden Progress in Deep Learning: SGD Learns Parities Near the Computational Limit
Figure 3 for Hidden Progress in Deep Learning: SGD Learns Parities Near the Computational Limit
Figure 4 for Hidden Progress in Deep Learning: SGD Learns Parities Near the Computational Limit
Viaarxiv icon

Inductive Biases and Variable Creation in Self-Attention Mechanisms

Add code
Oct 19, 2021
Figure 1 for Inductive Biases and Variable Creation in Self-Attention Mechanisms
Figure 2 for Inductive Biases and Variable Creation in Self-Attention Mechanisms
Figure 3 for Inductive Biases and Variable Creation in Self-Attention Mechanisms
Figure 4 for Inductive Biases and Variable Creation in Self-Attention Mechanisms
Viaarxiv icon

SGD on Neural Networks Learns Functions of Increasing Complexity

Add code
May 28, 2019
Figure 1 for SGD on Neural Networks Learns Functions of Increasing Complexity
Figure 2 for SGD on Neural Networks Learns Functions of Increasing Complexity
Figure 3 for SGD on Neural Networks Learns Functions of Increasing Complexity
Figure 4 for SGD on Neural Networks Learns Functions of Increasing Complexity
Viaarxiv icon