Picture for Christopher Summerfield

Christopher Summerfield

Reward Model Interpretability via Optimal and Pessimal Tokens

Add code
Jun 08, 2025
Viaarxiv icon

HiBayES: A Hierarchical Bayesian Modeling Framework for AI Evaluation Statistics

Add code
May 08, 2025
Viaarxiv icon

Increasing happiness through conversations with artificial intelligence

Add code
Apr 02, 2025
Viaarxiv icon

Language Agents as Digital Representatives in Collective Decision-Making

Add code
Feb 13, 2025
Viaarxiv icon

Flexible task abstractions emerge in linear networks with fast and bounded units

Add code
Nov 06, 2024
Viaarxiv icon

Early learning of the optimal constant solution in neural networks and humans

Add code
Jun 25, 2024
Figure 1 for Early learning of the optimal constant solution in neural networks and humans
Figure 2 for Early learning of the optimal constant solution in neural networks and humans
Figure 3 for Early learning of the optimal constant solution in neural networks and humans
Figure 4 for Early learning of the optimal constant solution in neural networks and humans
Viaarxiv icon

Using deep reinforcement learning to promote sustainable human behaviour on a common pool resource problem

Add code
Apr 23, 2024
Figure 1 for Using deep reinforcement learning to promote sustainable human behaviour on a common pool resource problem
Figure 2 for Using deep reinforcement learning to promote sustainable human behaviour on a common pool resource problem
Figure 3 for Using deep reinforcement learning to promote sustainable human behaviour on a common pool resource problem
Figure 4 for Using deep reinforcement learning to promote sustainable human behaviour on a common pool resource problem
Viaarxiv icon

Regularised neural networks mimic human insight

Add code
Feb 22, 2023
Figure 1 for Regularised neural networks mimic human insight
Figure 2 for Regularised neural networks mimic human insight
Figure 3 for Regularised neural networks mimic human insight
Figure 4 for Regularised neural networks mimic human insight
Viaarxiv icon

Fine-tuning language models to find agreement among humans with diverse preferences

Add code
Nov 28, 2022
Figure 1 for Fine-tuning language models to find agreement among humans with diverse preferences
Figure 2 for Fine-tuning language models to find agreement among humans with diverse preferences
Figure 3 for Fine-tuning language models to find agreement among humans with diverse preferences
Figure 4 for Fine-tuning language models to find agreement among humans with diverse preferences
Viaarxiv icon

Beyond Bayes-optimality: meta-learning what you know you don't know

Add code
Oct 12, 2022
Figure 1 for Beyond Bayes-optimality: meta-learning what you know you don't know
Figure 2 for Beyond Bayes-optimality: meta-learning what you know you don't know
Figure 3 for Beyond Bayes-optimality: meta-learning what you know you don't know
Figure 4 for Beyond Bayes-optimality: meta-learning what you know you don't know
Viaarxiv icon