Alert button
Picture for Albert Gu

Albert Gu

Alert button

Caduceus: Bi-Directional Equivariant Long-Range DNA Sequence Modeling

Add code
Bookmark button
Alert button
Mar 05, 2024
Yair Schiff, Chia-Hsiang Kao, Aaron Gokaslan, Tri Dao, Albert Gu, Volodymyr Kuleshov

Figure 1 for Caduceus: Bi-Directional Equivariant Long-Range DNA Sequence Modeling
Figure 2 for Caduceus: Bi-Directional Equivariant Long-Range DNA Sequence Modeling
Figure 3 for Caduceus: Bi-Directional Equivariant Long-Range DNA Sequence Modeling
Figure 4 for Caduceus: Bi-Directional Equivariant Long-Range DNA Sequence Modeling
Viaarxiv icon

Griffin: Mixing Gated Linear Recurrences with Local Attention for Efficient Language Models

Add code
Bookmark button
Alert button
Feb 29, 2024
Soham De, Samuel L. Smith, Anushan Fernando, Aleksandar Botev, George Cristian-Muraru, Albert Gu, Ruba Haroun, Leonard Berrada, Yutian Chen, Srivatsan Srinivasan, Guillaume Desjardins, Arnaud Doucet, David Budden, Yee Whye Teh, Razvan Pascanu, Nando De Freitas, Caglar Gulcehre

Viaarxiv icon

Mamba: Linear-Time Sequence Modeling with Selective State Spaces

Add code
Bookmark button
Alert button
Dec 01, 2023
Albert Gu, Tri Dao

Figure 1 for Mamba: Linear-Time Sequence Modeling with Selective State Spaces
Figure 2 for Mamba: Linear-Time Sequence Modeling with Selective State Spaces
Figure 3 for Mamba: Linear-Time Sequence Modeling with Selective State Spaces
Figure 4 for Mamba: Linear-Time Sequence Modeling with Selective State Spaces
Viaarxiv icon

Augmenting conformers with structured state space models for online speech recognition

Add code
Bookmark button
Alert button
Sep 15, 2023
Haozhe Shan, Albert Gu, Zhong Meng, Weiran Wang, Krzysztof Choromanski, Tara Sainath

Figure 1 for Augmenting conformers with structured state space models for online speech recognition
Figure 2 for Augmenting conformers with structured state space models for online speech recognition
Figure 3 for Augmenting conformers with structured state space models for online speech recognition
Figure 4 for Augmenting conformers with structured state space models for online speech recognition
Viaarxiv icon

Resurrecting Recurrent Neural Networks for Long Sequences

Add code
Bookmark button
Alert button
Mar 11, 2023
Antonio Orvieto, Samuel L Smith, Albert Gu, Anushan Fernando, Caglar Gulcehre, Razvan Pascanu, Soham De

Figure 1 for Resurrecting Recurrent Neural Networks for Long Sequences
Figure 2 for Resurrecting Recurrent Neural Networks for Long Sequences
Figure 3 for Resurrecting Recurrent Neural Networks for Long Sequences
Figure 4 for Resurrecting Recurrent Neural Networks for Long Sequences
Viaarxiv icon

Structured State Space Models for In-Context Reinforcement Learning

Add code
Bookmark button
Alert button
Mar 09, 2023
Chris Lu, Yannick Schroecker, Albert Gu, Emilio Parisotto, Jakob Foerster, Satinder Singh, Feryal Behbahani

Figure 1 for Structured State Space Models for In-Context Reinforcement Learning
Figure 2 for Structured State Space Models for In-Context Reinforcement Learning
Figure 3 for Structured State Space Models for In-Context Reinforcement Learning
Figure 4 for Structured State Space Models for In-Context Reinforcement Learning
Viaarxiv icon

Modelling Long Range Dependencies in N-D: From Task-Specific to a General Purpose CNN

Add code
Bookmark button
Alert button
Jan 25, 2023
David M. Knigge, David W. Romero, Albert Gu, Efstratios Gavves, Erik J. Bekkers, Jakub M. Tomczak, Mark Hoogendoorn, Jan-Jakob Sonke

Figure 1 for Modelling Long Range Dependencies in N-D: From Task-Specific to a General Purpose CNN
Figure 2 for Modelling Long Range Dependencies in N-D: From Task-Specific to a General Purpose CNN
Figure 3 for Modelling Long Range Dependencies in N-D: From Task-Specific to a General Purpose CNN
Figure 4 for Modelling Long Range Dependencies in N-D: From Task-Specific to a General Purpose CNN
Viaarxiv icon

Pretraining Without Attention

Add code
Bookmark button
Alert button
Dec 20, 2022
Junxiong Wang, Jing Nathan Yan, Albert Gu, Alexander M. Rush

Figure 1 for Pretraining Without Attention
Figure 2 for Pretraining Without Attention
Figure 3 for Pretraining Without Attention
Figure 4 for Pretraining Without Attention
Viaarxiv icon

S4ND: Modeling Images and Videos as Multidimensional Signals Using State Spaces

Add code
Bookmark button
Alert button
Oct 14, 2022
Eric Nguyen, Karan Goel, Albert Gu, Gordon W. Downs, Preey Shah, Tri Dao, Stephen A. Baccus, Christopher Ré

Figure 1 for S4ND: Modeling Images and Videos as Multidimensional Signals Using State Spaces
Figure 2 for S4ND: Modeling Images and Videos as Multidimensional Signals Using State Spaces
Figure 3 for S4ND: Modeling Images and Videos as Multidimensional Signals Using State Spaces
Figure 4 for S4ND: Modeling Images and Videos as Multidimensional Signals Using State Spaces
Viaarxiv icon

How to Train Your HiPPO: State Space Models with Generalized Orthogonal Basis Projections

Add code
Bookmark button
Alert button
Jun 24, 2022
Albert Gu, Isys Johnson, Aman Timalsina, Atri Rudra, Christopher Ré

Figure 1 for How to Train Your HiPPO: State Space Models with Generalized Orthogonal Basis Projections
Figure 2 for How to Train Your HiPPO: State Space Models with Generalized Orthogonal Basis Projections
Figure 3 for How to Train Your HiPPO: State Space Models with Generalized Orthogonal Basis Projections
Figure 4 for How to Train Your HiPPO: State Space Models with Generalized Orthogonal Basis Projections
Viaarxiv icon