Alert button
Picture for William Schuler

William Schuler

Alert button

Frequency Explains the Inverse Correlation of Large Language Models' Size, Training Data Amount, and Surprisal's Fit to Reading Times

Add code
Bookmark button
Alert button
Feb 03, 2024
Byung-Doh Oh, Shisen Yue, William Schuler

Viaarxiv icon

Token-wise Decomposition of Autoregressive Language Model Hidden States for Analyzing Model Predictions

Add code
Bookmark button
Alert button
May 17, 2023
Byung-Doh Oh, William Schuler

Figure 1 for Token-wise Decomposition of Autoregressive Language Model Hidden States for Analyzing Model Predictions
Figure 2 for Token-wise Decomposition of Autoregressive Language Model Hidden States for Analyzing Model Predictions
Figure 3 for Token-wise Decomposition of Autoregressive Language Model Hidden States for Analyzing Model Predictions
Figure 4 for Token-wise Decomposition of Autoregressive Language Model Hidden States for Analyzing Model Predictions
Viaarxiv icon

Transformer-Based LM Surprisal Predicts Human Reading Times Best with About Two Billion Training Tokens

Add code
Bookmark button
Alert button
Apr 22, 2023
Byung-Doh Oh, William Schuler

Figure 1 for Transformer-Based LM Surprisal Predicts Human Reading Times Best with About Two Billion Training Tokens
Figure 2 for Transformer-Based LM Surprisal Predicts Human Reading Times Best with About Two Billion Training Tokens
Figure 3 for Transformer-Based LM Surprisal Predicts Human Reading Times Best with About Two Billion Training Tokens
Figure 4 for Transformer-Based LM Surprisal Predicts Human Reading Times Best with About Two Billion Training Tokens
Viaarxiv icon

Why Does Surprisal From Larger Transformer-Based Language Models Provide a Poorer Fit to Human Reading Times?

Add code
Bookmark button
Alert button
Dec 23, 2022
Byung-Doh Oh, William Schuler

Figure 1 for Why Does Surprisal From Larger Transformer-Based Language Models Provide a Poorer Fit to Human Reading Times?
Figure 2 for Why Does Surprisal From Larger Transformer-Based Language Models Provide a Poorer Fit to Human Reading Times?
Figure 3 for Why Does Surprisal From Larger Transformer-Based Language Models Provide a Poorer Fit to Human Reading Times?
Figure 4 for Why Does Surprisal From Larger Transformer-Based Language Models Provide a Poorer Fit to Human Reading Times?
Viaarxiv icon

Entropy- and Distance-Based Predictors From GPT-2 Attention Patterns Predict Reading Times Over and Above GPT-2 Surprisal

Add code
Bookmark button
Alert button
Dec 21, 2022
Byung-Doh Oh, William Schuler

Figure 1 for Entropy- and Distance-Based Predictors From GPT-2 Attention Patterns Predict Reading Times Over and Above GPT-2 Surprisal
Figure 2 for Entropy- and Distance-Based Predictors From GPT-2 Attention Patterns Predict Reading Times Over and Above GPT-2 Surprisal
Figure 3 for Entropy- and Distance-Based Predictors From GPT-2 Attention Patterns Predict Reading Times Over and Above GPT-2 Surprisal
Figure 4 for Entropy- and Distance-Based Predictors From GPT-2 Attention Patterns Predict Reading Times Over and Above GPT-2 Surprisal
Viaarxiv icon

A Deep Learning Approach to Analyzing Continuous-Time Systems

Add code
Bookmark button
Alert button
Sep 25, 2022
Cory Shain, William Schuler

Figure 1 for A Deep Learning Approach to Analyzing Continuous-Time Systems
Figure 2 for A Deep Learning Approach to Analyzing Continuous-Time Systems
Figure 3 for A Deep Learning Approach to Analyzing Continuous-Time Systems
Figure 4 for A Deep Learning Approach to Analyzing Continuous-Time Systems
Viaarxiv icon

The Importance of Category Labels in Grammar Induction with Child-directed Utterances

Add code
Bookmark button
Alert button
Jun 20, 2020
Lifeng Jin, William Schuler

Figure 1 for The Importance of Category Labels in Grammar Induction with Child-directed Utterances
Figure 2 for The Importance of Category Labels in Grammar Induction with Child-directed Utterances
Figure 3 for The Importance of Category Labels in Grammar Induction with Child-directed Utterances
Figure 4 for The Importance of Category Labels in Grammar Induction with Child-directed Utterances
Viaarxiv icon

Depth-bounding is effective: Improvements and evaluation of unsupervised PCFG induction

Add code
Bookmark button
Alert button
Sep 10, 2018
Lifeng Jin, Finale Doshi-Velez, Timothy Miller, William Schuler, Lane Schwartz

Figure 1 for Depth-bounding is effective: Improvements and evaluation of unsupervised PCFG induction
Figure 2 for Depth-bounding is effective: Improvements and evaluation of unsupervised PCFG induction
Figure 3 for Depth-bounding is effective: Improvements and evaluation of unsupervised PCFG induction
Figure 4 for Depth-bounding is effective: Improvements and evaluation of unsupervised PCFG induction
Viaarxiv icon

Unsupervised Grammar Induction with Depth-bounded PCFG

Add code
Bookmark button
Alert button
Feb 26, 2018
Lifeng Jin, Finale Doshi-Velez, Timothy Miller, William Schuler, Lane Schwartz

Viaarxiv icon

Interleaved semantic interpretation in environment-based parsing

Add code
Bookmark button
Alert button
Jun 18, 2002
William Schuler

Figure 1 for Interleaved semantic interpretation in environment-based parsing
Figure 2 for Interleaved semantic interpretation in environment-based parsing
Figure 3 for Interleaved semantic interpretation in environment-based parsing
Viaarxiv icon