Alibi


What DINO saw: ALiBi positional encoding reduces positional bias in Vision Transformers

Add code
Mar 17, 2026
Viaarxiv icon

Surgical Repair of Collapsed Attention Heads in ALiBi Transformers

Add code
Mar 10, 2026
Viaarxiv icon

Position Encoding with Random Float Sampling Enhances Length Generalization of Transformers

Add code
Feb 15, 2026
Viaarxiv icon

Mitigating Position-Shift Failures in Text-Based Modular Arithmetic via Position Curriculum and Template Diversity

Add code
Jan 07, 2026
Viaarxiv icon

Group Representational Position Encoding

Add code
Dec 08, 2025
Figure 1 for Group Representational Position Encoding
Figure 2 for Group Representational Position Encoding
Figure 3 for Group Representational Position Encoding
Figure 4 for Group Representational Position Encoding
Viaarxiv icon

HoPE: Hyperbolic Rotary Positional Encoding for Stable Long-Range Dependency Modeling in Large Language Models

Add code
Sep 05, 2025
Viaarxiv icon

A standard transformer and attention with linear biases for molecular conformer generation

Add code
Jun 24, 2025
Viaarxiv icon

SeqPE: Transformer with Sequential Position Encoding

Add code
Jun 16, 2025
Viaarxiv icon

Theoretical Analysis of Positional Encodings in Transformer Models: Impact on Expressiveness and Generalization

Add code
Jun 05, 2025
Figure 1 for Theoretical Analysis of Positional Encodings in Transformer Models: Impact on Expressiveness and Generalization
Figure 2 for Theoretical Analysis of Positional Encodings in Transformer Models: Impact on Expressiveness and Generalization
Figure 3 for Theoretical Analysis of Positional Encodings in Transformer Models: Impact on Expressiveness and Generalization
Figure 4 for Theoretical Analysis of Positional Encodings in Transformer Models: Impact on Expressiveness and Generalization
Viaarxiv icon

Bayesian Attention Mechanism: A Probabilistic Framework for Positional Encoding and Context Length Extrapolation

Add code
May 28, 2025
Viaarxiv icon