Alert button
Picture for Che Zheng

Che Zheng

Alert button

StructFormer: Joint Unsupervised Induction of Dependency and Constituency Structure from Masked Language Modeling

Dec 15, 2020
Yikang Shen, Yi Tay, Che Zheng, Dara Bahri, Donald Metzler, Aaron Courville

Figure 1 for StructFormer: Joint Unsupervised Induction of Dependency and Constituency Structure from Masked Language Modeling
Figure 2 for StructFormer: Joint Unsupervised Induction of Dependency and Constituency Structure from Masked Language Modeling
Figure 3 for StructFormer: Joint Unsupervised Induction of Dependency and Constituency Structure from Masked Language Modeling
Figure 4 for StructFormer: Joint Unsupervised Induction of Dependency and Constituency Structure from Masked Language Modeling
Viaarxiv icon

Surprise: Result List Truncation via Extreme Value Theory

Oct 19, 2020
Dara Bahri, Che Zheng, Yi Tay, Donald Metzler, Andrew Tomkins

Figure 1 for Surprise: Result List Truncation via Extreme Value Theory
Figure 2 for Surprise: Result List Truncation via Extreme Value Theory
Figure 3 for Surprise: Result List Truncation via Extreme Value Theory
Figure 4 for Surprise: Result List Truncation via Extreme Value Theory
Viaarxiv icon

Generative Models are Unsupervised Predictors of Page Quality: A Colossal-Scale Study

Aug 17, 2020
Dara Bahri, Yi Tay, Che Zheng, Donald Metzler, Cliff Brunk, Andrew Tomkins

Figure 1 for Generative Models are Unsupervised Predictors of Page Quality: A Colossal-Scale Study
Figure 2 for Generative Models are Unsupervised Predictors of Page Quality: A Colossal-Scale Study
Figure 3 for Generative Models are Unsupervised Predictors of Page Quality: A Colossal-Scale Study
Figure 4 for Generative Models are Unsupervised Predictors of Page Quality: A Colossal-Scale Study
Viaarxiv icon

Synthesizer: Rethinking Self-Attention in Transformer Models

May 02, 2020
Yi Tay, Dara Bahri, Donald Metzler, Da-Cheng Juan, Zhe Zhao, Che Zheng

Figure 1 for Synthesizer: Rethinking Self-Attention in Transformer Models
Figure 2 for Synthesizer: Rethinking Self-Attention in Transformer Models
Figure 3 for Synthesizer: Rethinking Self-Attention in Transformer Models
Figure 4 for Synthesizer: Rethinking Self-Attention in Transformer Models
Viaarxiv icon

Choppy: Cut Transformer For Ranked List Truncation

Apr 26, 2020
Dara Bahri, Yi Tay, Che Zheng, Donald Metzler, Andrew Tomkins

Figure 1 for Choppy: Cut Transformer For Ranked List Truncation
Figure 2 for Choppy: Cut Transformer For Ranked List Truncation
Figure 3 for Choppy: Cut Transformer For Ranked List Truncation
Viaarxiv icon

Reverse Engineering Configurations of Neural Text Generation Models

Apr 13, 2020
Yi Tay, Dara Bahri, Che Zheng, Clifford Brunk, Donald Metzler, Andrew Tomkins

Figure 1 for Reverse Engineering Configurations of Neural Text Generation Models
Figure 2 for Reverse Engineering Configurations of Neural Text Generation Models
Figure 3 for Reverse Engineering Configurations of Neural Text Generation Models
Viaarxiv icon