Alert button
Picture for Ta-Chung Chi

Ta-Chung Chi

Alert button

Attention Alignment and Flexible Positional Embeddings Improve Transformer Length Extrapolation

Nov 15, 2023
Ta-Chung Chi, Ting-Han Fan, Alexander I. Rudnicky

Viaarxiv icon

Advancing Regular Language Reasoning in Linear Recurrent Neural Networks

Sep 14, 2023
Ting-Han Fan, Ta-Chung Chi, Alexander I. Rudnicky

Viaarxiv icon

Structured Dialogue Discourse Parsing

Jun 26, 2023
Ta-Chung Chi, Alexander I. Rudnicky

Figure 1 for Structured Dialogue Discourse Parsing
Figure 2 for Structured Dialogue Discourse Parsing
Figure 3 for Structured Dialogue Discourse Parsing
Figure 4 for Structured Dialogue Discourse Parsing
Viaarxiv icon

PESCO: Prompt-enhanced Self Contrastive Learning for Zero-shot Text Classification

May 24, 2023
Yau-Shian Wang, Ta-Chung Chi, Ruohong Zhang, Yiming Yang

Figure 1 for PESCO: Prompt-enhanced Self Contrastive Learning for Zero-shot Text Classification
Figure 2 for PESCO: Prompt-enhanced Self Contrastive Learning for Zero-shot Text Classification
Figure 3 for PESCO: Prompt-enhanced Self Contrastive Learning for Zero-shot Text Classification
Figure 4 for PESCO: Prompt-enhanced Self Contrastive Learning for Zero-shot Text Classification
Viaarxiv icon

Latent Positional Information is in the Self-Attention Variance of Transformer Language Models Without Positional Embeddings

May 23, 2023
Ta-Chung Chi, Ting-Han Fan, Li-Wei Chen, Alexander I. Rudnicky, Peter J. Ramadge

Figure 1 for Latent Positional Information is in the Self-Attention Variance of Transformer Language Models Without Positional Embeddings
Figure 2 for Latent Positional Information is in the Self-Attention Variance of Transformer Language Models Without Positional Embeddings
Figure 3 for Latent Positional Information is in the Self-Attention Variance of Transformer Language Models Without Positional Embeddings
Figure 4 for Latent Positional Information is in the Self-Attention Variance of Transformer Language Models Without Positional Embeddings
Viaarxiv icon

Transformer Working Memory Enables Regular Language Reasoning and Natural Language Length Extrapolation

May 05, 2023
Ta-Chung Chi, Ting-Han Fan, Alexander I. Rudnicky, Peter J. Ramadge

Figure 1 for Transformer Working Memory Enables Regular Language Reasoning and Natural Language Length Extrapolation
Figure 2 for Transformer Working Memory Enables Regular Language Reasoning and Natural Language Length Extrapolation
Figure 3 for Transformer Working Memory Enables Regular Language Reasoning and Natural Language Length Extrapolation
Figure 4 for Transformer Working Memory Enables Regular Language Reasoning and Natural Language Length Extrapolation
Viaarxiv icon

Receptive Field Alignment Enables Transformer Length Extrapolation

Dec 20, 2022
Ta-Chung Chi, Ting-Han Fan, Alexander I. Rudnicky

Figure 1 for Receptive Field Alignment Enables Transformer Length Extrapolation
Figure 2 for Receptive Field Alignment Enables Transformer Length Extrapolation
Figure 3 for Receptive Field Alignment Enables Transformer Length Extrapolation
Figure 4 for Receptive Field Alignment Enables Transformer Length Extrapolation
Viaarxiv icon

On Task-Adaptive Pretraining for Dialogue Response Selection

Oct 08, 2022
Tzu-Hsiang Lin, Ta-Chung Chi, Anna Rumshisky

Figure 1 for On Task-Adaptive Pretraining for Dialogue Response Selection
Figure 2 for On Task-Adaptive Pretraining for Dialogue Response Selection
Figure 3 for On Task-Adaptive Pretraining for Dialogue Response Selection
Figure 4 for On Task-Adaptive Pretraining for Dialogue Response Selection
Viaarxiv icon

Training Discrete Deep Generative Models via Gapped Straight-Through Estimator

Jun 15, 2022
Ting-Han Fan, Ta-Chung Chi, Alexander I. Rudnicky, Peter J. Ramadge

Figure 1 for Training Discrete Deep Generative Models via Gapped Straight-Through Estimator
Figure 2 for Training Discrete Deep Generative Models via Gapped Straight-Through Estimator
Figure 3 for Training Discrete Deep Generative Models via Gapped Straight-Through Estimator
Figure 4 for Training Discrete Deep Generative Models via Gapped Straight-Through Estimator
Viaarxiv icon