Picture for Aaron Mueller

Aaron Mueller

Missed Causes and Ambiguous Effects: Counterfactuals Pose Challenges for Interpreting Neural Networks

Add code
Jul 05, 2024
Viaarxiv icon

[Call for Papers] The 2nd BabyLM Challenge: Sample-efficient pretraining on a developmentally plausible corpus

Add code
Apr 09, 2024
Figure 1 for [Call for Papers] The 2nd BabyLM Challenge: Sample-efficient pretraining on a developmentally plausible corpus
Viaarxiv icon

Sparse Feature Circuits: Discovering and Editing Interpretable Causal Graphs in Language Models

Add code
Mar 31, 2024
Figure 1 for Sparse Feature Circuits: Discovering and Editing Interpretable Causal Graphs in Language Models
Figure 2 for Sparse Feature Circuits: Discovering and Editing Interpretable Causal Graphs in Language Models
Figure 3 for Sparse Feature Circuits: Discovering and Editing Interpretable Causal Graphs in Language Models
Figure 4 for Sparse Feature Circuits: Discovering and Editing Interpretable Causal Graphs in Language Models
Viaarxiv icon

In-context Learning Generalizes, But Not Always Robustly: The Case of Syntax

Add code
Nov 13, 2023
Figure 1 for In-context Learning Generalizes, But Not Always Robustly: The Case of Syntax
Figure 2 for In-context Learning Generalizes, But Not Always Robustly: The Case of Syntax
Figure 3 for In-context Learning Generalizes, But Not Always Robustly: The Case of Syntax
Figure 4 for In-context Learning Generalizes, But Not Always Robustly: The Case of Syntax
Viaarxiv icon

Function Vectors in Large Language Models

Add code
Oct 23, 2023
Viaarxiv icon

Meta-training with Demonstration Retrieval for Efficient Few-shot Learning

Add code
Jun 30, 2023
Figure 1 for Meta-training with Demonstration Retrieval for Efficient Few-shot Learning
Figure 2 for Meta-training with Demonstration Retrieval for Efficient Few-shot Learning
Figure 3 for Meta-training with Demonstration Retrieval for Efficient Few-shot Learning
Figure 4 for Meta-training with Demonstration Retrieval for Efficient Few-shot Learning
Viaarxiv icon

Inverse Scaling: When Bigger Isn't Better

Add code
Jun 15, 2023
Figure 1 for Inverse Scaling: When Bigger Isn't Better
Figure 2 for Inverse Scaling: When Bigger Isn't Better
Figure 3 for Inverse Scaling: When Bigger Isn't Better
Figure 4 for Inverse Scaling: When Bigger Isn't Better
Viaarxiv icon

How to Plant Trees in Language Models: Data and Architectural Effects on the Emergence of Syntactic Inductive Biases

Add code
May 31, 2023
Figure 1 for How to Plant Trees in Language Models: Data and Architectural Effects on the Emergence of Syntactic Inductive Biases
Figure 2 for How to Plant Trees in Language Models: Data and Architectural Effects on the Emergence of Syntactic Inductive Biases
Figure 3 for How to Plant Trees in Language Models: Data and Architectural Effects on the Emergence of Syntactic Inductive Biases
Figure 4 for How to Plant Trees in Language Models: Data and Architectural Effects on the Emergence of Syntactic Inductive Biases
Viaarxiv icon

Call for Papers -- The BabyLM Challenge: Sample-efficient pretraining on a developmentally plausible corpus

Add code
Jan 27, 2023
Figure 1 for Call for Papers -- The BabyLM Challenge: Sample-efficient pretraining on a developmentally plausible corpus
Figure 2 for Call for Papers -- The BabyLM Challenge: Sample-efficient pretraining on a developmentally plausible corpus
Viaarxiv icon

Language model acceptability judgements are not always robust to context

Add code
Dec 18, 2022
Figure 1 for Language model acceptability judgements are not always robust to context
Figure 2 for Language model acceptability judgements are not always robust to context
Figure 3 for Language model acceptability judgements are not always robust to context
Figure 4 for Language model acceptability judgements are not always robust to context
Viaarxiv icon