Picture for Ryan Cotterell

Ryan Cotterell

ETH Zurich

Characterizing the Expressivity of Transformer Language Models

Add code
May 29, 2025
Figure 1 for Characterizing the Expressivity of Transformer Language Models
Figure 2 for Characterizing the Expressivity of Transformer Language Models
Figure 3 for Characterizing the Expressivity of Transformer Language Models
Figure 4 for Characterizing the Expressivity of Transformer Language Models
Viaarxiv icon

Towards Developmentally Plausible Rewards: Communicative Success as a Learning Signal for Interactive Language Models

Add code
May 09, 2025
Figure 1 for Towards Developmentally Plausible Rewards: Communicative Success as a Learning Signal for Interactive Language Models
Figure 2 for Towards Developmentally Plausible Rewards: Communicative Success as a Learning Signal for Interactive Language Models
Figure 3 for Towards Developmentally Plausible Rewards: Communicative Success as a Learning Signal for Interactive Language Models
Figure 4 for Towards Developmentally Plausible Rewards: Communicative Success as a Learning Signal for Interactive Language Models
Viaarxiv icon

Syntactic and Semantic Control of Large Language Models via Sequential Monte Carlo

Add code
Apr 18, 2025
Figure 1 for Syntactic and Semantic Control of Large Language Models via Sequential Monte Carlo
Figure 2 for Syntactic and Semantic Control of Large Language Models via Sequential Monte Carlo
Figure 3 for Syntactic and Semantic Control of Large Language Models via Sequential Monte Carlo
Figure 4 for Syntactic and Semantic Control of Large Language Models via Sequential Monte Carlo
Viaarxiv icon

Better Estimation of the KL Divergence Between Language Models

Add code
Apr 14, 2025
Viaarxiv icon

Findings of the BabyLM Challenge: Sample-Efficient Pretraining on Developmentally Plausible Corpora

Add code
Apr 10, 2025
Figure 1 for Findings of the BabyLM Challenge: Sample-Efficient Pretraining on Developmentally Plausible Corpora
Figure 2 for Findings of the BabyLM Challenge: Sample-Efficient Pretraining on Developmentally Plausible Corpora
Figure 3 for Findings of the BabyLM Challenge: Sample-Efficient Pretraining on Developmentally Plausible Corpora
Figure 4 for Findings of the BabyLM Challenge: Sample-Efficient Pretraining on Developmentally Plausible Corpora
Viaarxiv icon

Fast Controlled Generation from Language Models with Adaptive Weighted Rejection Sampling

Add code
Apr 07, 2025
Viaarxiv icon

Taxonomy-Aware Evaluation of Vision-Language Models

Add code
Apr 07, 2025
Figure 1 for Taxonomy-Aware Evaluation of Vision-Language Models
Viaarxiv icon

Unique Hard Attention: A Tale of Two Sides

Add code
Mar 18, 2025
Figure 1 for Unique Hard Attention: A Tale of Two Sides
Figure 2 for Unique Hard Attention: A Tale of Two Sides
Viaarxiv icon

Can Language Models Learn Typologically Implausible Languages?

Add code
Feb 17, 2025
Viaarxiv icon

A Distributional Perspective on Word Learning in Neural Language Models

Add code
Feb 09, 2025
Viaarxiv icon