Alert button
Picture for Shanda Li

Shanda Li

Alert button

Functional Interpolation for Relative Positions Improves Long Context Transformers

Add code
Bookmark button
Alert button
Oct 06, 2023
Shanda Li, Chong You, Guru Guruganesh, Joshua Ainslie, Santiago Ontanon, Manzil Zaheer, Sumit Sanghai, Yiming Yang, Sanjiv Kumar, Srinadh Bhojanapalli

Viaarxiv icon

Learning a Fourier Transform for Linear Relative Positional Encodings in Transformers

Add code
Bookmark button
Alert button
Feb 03, 2023
Krzysztof Marcin Choromanski, Shanda Li, Valerii Likhosherstov, Kumar Avinava Dubey, Shengjie Luo, Di He, Yiming Yang, Tamas Sarlos, Thomas Weingarten, Adrian Weller

Figure 1 for Learning a Fourier Transform for Linear Relative Positional Encodings in Transformers
Figure 2 for Learning a Fourier Transform for Linear Relative Positional Encodings in Transformers
Figure 3 for Learning a Fourier Transform for Linear Relative Positional Encodings in Transformers
Figure 4 for Learning a Fourier Transform for Linear Relative Positional Encodings in Transformers
Viaarxiv icon

Is $L^2$ Physics-Informed Loss Always Suitable for Training Physics-Informed Neural Network?

Add code
Bookmark button
Alert button
Jun 04, 2022
Chuwei Wang, Shanda Li, Di He, Liwei Wang

Figure 1 for Is $L^2$ Physics-Informed Loss Always Suitable for Training Physics-Informed Neural Network?
Figure 2 for Is $L^2$ Physics-Informed Loss Always Suitable for Training Physics-Informed Neural Network?
Figure 3 for Is $L^2$ Physics-Informed Loss Always Suitable for Training Physics-Informed Neural Network?
Figure 4 for Is $L^2$ Physics-Informed Loss Always Suitable for Training Physics-Informed Neural Network?
Viaarxiv icon

Your Transformer May Not be as Powerful as You Expect

Add code
Bookmark button
Alert button
May 26, 2022
Shengjie Luo, Shanda Li, Shuxin Zheng, Tie-Yan Liu, Liwei Wang, Di He

Figure 1 for Your Transformer May Not be as Powerful as You Expect
Figure 2 for Your Transformer May Not be as Powerful as You Expect
Figure 3 for Your Transformer May Not be as Powerful as You Expect
Figure 4 for Your Transformer May Not be as Powerful as You Expect
Viaarxiv icon

Learning Physics-Informed Neural Networks without Stacked Back-propagation

Add code
Bookmark button
Alert button
Feb 18, 2022
Di He, Wenlei Shi, Shanda Li, Xiaotian Gao, Jia Zhang, Jiang Bian, Liwei Wang, Tie-Yan Liu

Figure 1 for Learning Physics-Informed Neural Networks without Stacked Back-propagation
Figure 2 for Learning Physics-Informed Neural Networks without Stacked Back-propagation
Figure 3 for Learning Physics-Informed Neural Networks without Stacked Back-propagation
Figure 4 for Learning Physics-Informed Neural Networks without Stacked Back-propagation
Viaarxiv icon

Can Vision Transformers Perform Convolution?

Add code
Bookmark button
Alert button
Nov 03, 2021
Shanda Li, Xiangning Chen, Di He, Cho-Jui Hsieh

Figure 1 for Can Vision Transformers Perform Convolution?
Figure 2 for Can Vision Transformers Perform Convolution?
Figure 3 for Can Vision Transformers Perform Convolution?
Viaarxiv icon

Stable, Fast and Accurate: Kernelized Attention with Relative Positional Encoding

Add code
Bookmark button
Alert button
Jun 23, 2021
Shengjie Luo, Shanda Li, Tianle Cai, Di He, Dinglan Peng, Shuxin Zheng, Guolin Ke, Liwei Wang, Tie-Yan Liu

Figure 1 for Stable, Fast and Accurate: Kernelized Attention with Relative Positional Encoding
Figure 2 for Stable, Fast and Accurate: Kernelized Attention with Relative Positional Encoding
Figure 3 for Stable, Fast and Accurate: Kernelized Attention with Relative Positional Encoding
Figure 4 for Stable, Fast and Accurate: Kernelized Attention with Relative Positional Encoding
Viaarxiv icon