Alert button
Picture for Dara Bahri

Dara Bahri

Alert button

Surprise: Result List Truncation via Extreme Value Theory

Add code
Bookmark button
Alert button
Oct 19, 2020
Dara Bahri, Che Zheng, Yi Tay, Donald Metzler, Andrew Tomkins

Figure 1 for Surprise: Result List Truncation via Extreme Value Theory
Figure 2 for Surprise: Result List Truncation via Extreme Value Theory
Figure 3 for Surprise: Result List Truncation via Extreme Value Theory
Figure 4 for Surprise: Result List Truncation via Extreme Value Theory
Viaarxiv icon

Efficient Transformers: A Survey

Add code
Bookmark button
Alert button
Sep 16, 2020
Yi Tay, Mostafa Dehghani, Dara Bahri, Donald Metzler

Figure 1 for Efficient Transformers: A Survey
Figure 2 for Efficient Transformers: A Survey
Figure 3 for Efficient Transformers: A Survey
Figure 4 for Efficient Transformers: A Survey
Viaarxiv icon

Generative Models are Unsupervised Predictors of Page Quality: A Colossal-Scale Study

Add code
Bookmark button
Alert button
Aug 17, 2020
Dara Bahri, Yi Tay, Che Zheng, Donald Metzler, Cliff Brunk, Andrew Tomkins

Figure 1 for Generative Models are Unsupervised Predictors of Page Quality: A Colossal-Scale Study
Figure 2 for Generative Models are Unsupervised Predictors of Page Quality: A Colossal-Scale Study
Figure 3 for Generative Models are Unsupervised Predictors of Page Quality: A Colossal-Scale Study
Figure 4 for Generative Models are Unsupervised Predictors of Page Quality: A Colossal-Scale Study
Viaarxiv icon

HyperGrid: Efficient Multi-Task Transformers with Grid-wise Decomposable Hyper Projections

Add code
Bookmark button
Alert button
Jul 12, 2020
Yi Tay, Zhe Zhao, Dara Bahri, Donald Metzler, Da-Cheng Juan

Figure 1 for HyperGrid: Efficient Multi-Task Transformers with Grid-wise Decomposable Hyper Projections
Figure 2 for HyperGrid: Efficient Multi-Task Transformers with Grid-wise Decomposable Hyper Projections
Figure 3 for HyperGrid: Efficient Multi-Task Transformers with Grid-wise Decomposable Hyper Projections
Figure 4 for HyperGrid: Efficient Multi-Task Transformers with Grid-wise Decomposable Hyper Projections
Viaarxiv icon

Synthesizer: Rethinking Self-Attention in Transformer Models

Add code
Bookmark button
Alert button
May 02, 2020
Yi Tay, Dara Bahri, Donald Metzler, Da-Cheng Juan, Zhe Zhao, Che Zheng

Figure 1 for Synthesizer: Rethinking Self-Attention in Transformer Models
Figure 2 for Synthesizer: Rethinking Self-Attention in Transformer Models
Figure 3 for Synthesizer: Rethinking Self-Attention in Transformer Models
Figure 4 for Synthesizer: Rethinking Self-Attention in Transformer Models
Viaarxiv icon

Deep k-NN for Noisy Labels

Add code
Bookmark button
Alert button
Apr 26, 2020
Dara Bahri, Heinrich Jiang, Maya Gupta

Figure 1 for Deep k-NN for Noisy Labels
Figure 2 for Deep k-NN for Noisy Labels
Figure 3 for Deep k-NN for Noisy Labels
Figure 4 for Deep k-NN for Noisy Labels
Viaarxiv icon

Choppy: Cut Transformer For Ranked List Truncation

Add code
Bookmark button
Alert button
Apr 26, 2020
Dara Bahri, Yi Tay, Che Zheng, Donald Metzler, Andrew Tomkins

Figure 1 for Choppy: Cut Transformer For Ranked List Truncation
Figure 2 for Choppy: Cut Transformer For Ranked List Truncation
Figure 3 for Choppy: Cut Transformer For Ranked List Truncation
Viaarxiv icon

Reverse Engineering Configurations of Neural Text Generation Models

Add code
Bookmark button
Alert button
Apr 13, 2020
Yi Tay, Dara Bahri, Che Zheng, Clifford Brunk, Donald Metzler, Andrew Tomkins

Figure 1 for Reverse Engineering Configurations of Neural Text Generation Models
Figure 2 for Reverse Engineering Configurations of Neural Text Generation Models
Figure 3 for Reverse Engineering Configurations of Neural Text Generation Models
Viaarxiv icon

Sparse Sinkhorn Attention

Add code
Bookmark button
Alert button
Feb 26, 2020
Yi Tay, Dara Bahri, Liu Yang, Donald Metzler, Da-Cheng Juan

Figure 1 for Sparse Sinkhorn Attention
Figure 2 for Sparse Sinkhorn Attention
Figure 3 for Sparse Sinkhorn Attention
Figure 4 for Sparse Sinkhorn Attention
Viaarxiv icon