Alert button
Picture for Yin-Wen Chang

Yin-Wen Chang

Alert button

Leveraging redundancy in attention with Reuse Transformers

Add code
Bookmark button
Alert button
Oct 13, 2021
Srinadh Bhojanapalli, Ayan Chakrabarti, Andreas Veit, Michal Lukasik, Himanshu Jain, Frederick Liu, Yin-Wen Chang, Sanjiv Kumar

Figure 1 for Leveraging redundancy in attention with Reuse Transformers
Figure 2 for Leveraging redundancy in attention with Reuse Transformers
Figure 3 for Leveraging redundancy in attention with Reuse Transformers
Figure 4 for Leveraging redundancy in attention with Reuse Transformers
Viaarxiv icon

Demystifying the Better Performance of Position Encoding Variants for Transformer

Add code
Bookmark button
Alert button
Apr 18, 2021
Pu-Chin Chen, Henry Tsai, Srinadh Bhojanapalli, Hyung Won Chung, Yin-Wen Chang, Chun-Sung Ferng

Figure 1 for Demystifying the Better Performance of Position Encoding Variants for Transformer
Figure 2 for Demystifying the Better Performance of Position Encoding Variants for Transformer
Figure 3 for Demystifying the Better Performance of Position Encoding Variants for Transformer
Figure 4 for Demystifying the Better Performance of Position Encoding Variants for Transformer
Viaarxiv icon

$O(n)$ Connections are Expressive Enough: Universal Approximability of Sparse Transformers

Add code
Bookmark button
Alert button
Jun 08, 2020
Chulhee Yun, Yin-Wen Chang, Srinadh Bhojanapalli, Ankit Singh Rawat, Sashank J. Reddi, Sanjiv Kumar

Figure 1 for $O(n)$ Connections are Expressive Enough: Universal Approximability of Sparse Transformers
Figure 2 for $O(n)$ Connections are Expressive Enough: Universal Approximability of Sparse Transformers
Viaarxiv icon

Pre-training Tasks for Embedding-based Large-scale Retrieval

Add code
Bookmark button
Alert button
Feb 10, 2020
Wei-Cheng Chang, Felix X. Yu, Yin-Wen Chang, Yiming Yang, Sanjiv Kumar

Figure 1 for Pre-training Tasks for Embedding-based Large-scale Retrieval
Figure 2 for Pre-training Tasks for Embedding-based Large-scale Retrieval
Figure 3 for Pre-training Tasks for Embedding-based Large-scale Retrieval
Figure 4 for Pre-training Tasks for Embedding-based Large-scale Retrieval
Viaarxiv icon