Picture for Goro Kobayashi

Goro Kobayashi

PLaMo 2 Technical Report

Add code
Sep 05, 2025
Viaarxiv icon

Can Input Attributions Interpret the Inductive Reasoning Process Elicited in In-Context Learning?

Add code
Dec 20, 2024
Figure 1 for Can Input Attributions Interpret the Inductive Reasoning Process Elicited in In-Context Learning?
Figure 2 for Can Input Attributions Interpret the Inductive Reasoning Process Elicited in In-Context Learning?
Figure 3 for Can Input Attributions Interpret the Inductive Reasoning Process Elicited in In-Context Learning?
Figure 4 for Can Input Attributions Interpret the Inductive Reasoning Process Elicited in In-Context Learning?
Viaarxiv icon

Contrastive Learning-based Sentence Encoders Implicitly Weight Informative Words

Add code
Oct 24, 2023
Viaarxiv icon

Assessing Step-by-Step Reasoning against Lexical Negation: A Case Study on Syllogism

Add code
Oct 23, 2023
Figure 1 for Assessing Step-by-Step Reasoning against Lexical Negation: A Case Study on Syllogism
Figure 2 for Assessing Step-by-Step Reasoning against Lexical Negation: A Case Study on Syllogism
Figure 3 for Assessing Step-by-Step Reasoning against Lexical Negation: A Case Study on Syllogism
Figure 4 for Assessing Step-by-Step Reasoning against Lexical Negation: A Case Study on Syllogism
Viaarxiv icon

Transformer Language Models Handle Word Frequency in Prediction Head

Add code
May 29, 2023
Figure 1 for Transformer Language Models Handle Word Frequency in Prediction Head
Figure 2 for Transformer Language Models Handle Word Frequency in Prediction Head
Figure 3 for Transformer Language Models Handle Word Frequency in Prediction Head
Figure 4 for Transformer Language Models Handle Word Frequency in Prediction Head
Viaarxiv icon

Feed-Forward Blocks Control Contextualization in Masked Language Models

Add code
Feb 01, 2023
Figure 1 for Feed-Forward Blocks Control Contextualization in Masked Language Models
Figure 2 for Feed-Forward Blocks Control Contextualization in Masked Language Models
Figure 3 for Feed-Forward Blocks Control Contextualization in Masked Language Models
Figure 4 for Feed-Forward Blocks Control Contextualization in Masked Language Models
Viaarxiv icon

Incorporating Residual and Normalization Layers into Analysis of Masked Language Models

Add code
Sep 15, 2021
Figure 1 for Incorporating Residual and Normalization Layers into Analysis of Masked Language Models
Figure 2 for Incorporating Residual and Normalization Layers into Analysis of Masked Language Models
Figure 3 for Incorporating Residual and Normalization Layers into Analysis of Masked Language Models
Figure 4 for Incorporating Residual and Normalization Layers into Analysis of Masked Language Models
Viaarxiv icon

Attention Module is Not Only a Weight: Analyzing Transformers with Vector Norms

Add code
Apr 21, 2020
Figure 1 for Attention Module is Not Only a Weight: Analyzing Transformers with Vector Norms
Figure 2 for Attention Module is Not Only a Weight: Analyzing Transformers with Vector Norms
Figure 3 for Attention Module is Not Only a Weight: Analyzing Transformers with Vector Norms
Figure 4 for Attention Module is Not Only a Weight: Analyzing Transformers with Vector Norms
Viaarxiv icon