Alert button
Picture for Ziyang Wu

Ziyang Wu

Alert button

LLoCO: Learning Long Contexts Offline

Add code
Bookmark button
Alert button
Apr 11, 2024
Sijun Tan, Xiuyu Li, Shishir Patil, Ziyang Wu, Tianjun Zhang, Kurt Keutzer, Joseph E. Gonzalez, Raluca Ada Popa

Viaarxiv icon

Masked Completion via Structured Diffusion with White-Box Transformers

Add code
Bookmark button
Alert button
Apr 03, 2024
Druv Pai, Ziyang Wu, Sam Buchanan, Yaodong Yu, Yi Ma

Viaarxiv icon

When Do We Not Need Larger Vision Models?

Add code
Bookmark button
Alert button
Mar 19, 2024
Baifeng Shi, Ziyang Wu, Maolin Mao, Xin Wang, Trevor Darrell

Figure 1 for When Do We Not Need Larger Vision Models?
Figure 2 for When Do We Not Need Larger Vision Models?
Figure 3 for When Do We Not Need Larger Vision Models?
Figure 4 for When Do We Not Need Larger Vision Models?
Viaarxiv icon

White-Box Transformers via Sparse Rate Reduction: Compression Is All There Is?

Add code
Bookmark button
Alert button
Nov 24, 2023
Yaodong Yu, Sam Buchanan, Druv Pai, Tianzhe Chu, Ziyang Wu, Shengbang Tong, Hao Bai, Yuexiang Zhai, Benjamin D. Haeffele, Yi Ma

Viaarxiv icon

Emergence of Segmentation with Minimalistic White-Box Transformers

Add code
Bookmark button
Alert button
Aug 30, 2023
Yaodong Yu, Tianzhe Chu, Shengbang Tong, Ziyang Wu, Druv Pai, Sam Buchanan, Yi Ma

Figure 1 for Emergence of Segmentation with Minimalistic White-Box Transformers
Figure 2 for Emergence of Segmentation with Minimalistic White-Box Transformers
Figure 3 for Emergence of Segmentation with Minimalistic White-Box Transformers
Figure 4 for Emergence of Segmentation with Minimalistic White-Box Transformers
Viaarxiv icon

White-Box Transformers via Sparse Rate Reduction

Add code
Bookmark button
Alert button
Jun 01, 2023
Yaodong Yu, Sam Buchanan, Druv Pai, Tianzhe Chu, Ziyang Wu, Shengbang Tong, Benjamin D. Haeffele, Yi Ma

Figure 1 for White-Box Transformers via Sparse Rate Reduction
Figure 2 for White-Box Transformers via Sparse Rate Reduction
Figure 3 for White-Box Transformers via Sparse Rate Reduction
Figure 4 for White-Box Transformers via Sparse Rate Reduction
Viaarxiv icon

Efficient Maximal Coding Rate Reduction by Variational Forms

Add code
Bookmark button
Alert button
Mar 31, 2022
Christina Baek, Ziyang Wu, Kwan Ho Ryan Chan, Tianjiao Ding, Yi Ma, Benjamin D. Haeffele

Figure 1 for Efficient Maximal Coding Rate Reduction by Variational Forms
Figure 2 for Efficient Maximal Coding Rate Reduction by Variational Forms
Figure 3 for Efficient Maximal Coding Rate Reduction by Variational Forms
Figure 4 for Efficient Maximal Coding Rate Reduction by Variational Forms
Viaarxiv icon

Incremental Learning of Structured Memory via Closed-Loop Transcription

Add code
Bookmark button
Alert button
Feb 14, 2022
Shengbang Tong, Xili Dai, Ziyang Wu, Mingyang Li, Brent Yi, Yi Ma

Figure 1 for Incremental Learning of Structured Memory via Closed-Loop Transcription
Figure 2 for Incremental Learning of Structured Memory via Closed-Loop Transcription
Figure 3 for Incremental Learning of Structured Memory via Closed-Loop Transcription
Figure 4 for Incremental Learning of Structured Memory via Closed-Loop Transcription
Viaarxiv icon

Closed-Loop Data Transcription to an LDR via Minimaxing Rate Reduction

Add code
Bookmark button
Alert button
Nov 12, 2021
Xili Dai, Shengbang Tong, Mingyang Li, Ziyang Wu, Kwan Ho Ryan Chan, Pengyuan Zhai, Yaodong Yu, Michael Psenka, Xiaojun Yuan, Heung Yeung Shum, Yi Ma

Figure 1 for Closed-Loop Data Transcription to an LDR via Minimaxing Rate Reduction
Figure 2 for Closed-Loop Data Transcription to an LDR via Minimaxing Rate Reduction
Figure 3 for Closed-Loop Data Transcription to an LDR via Minimaxing Rate Reduction
Figure 4 for Closed-Loop Data Transcription to an LDR via Minimaxing Rate Reduction
Viaarxiv icon

How Low Can We Go: Trading Memory for Error in Low-Precision Training

Add code
Bookmark button
Alert button
Jun 18, 2021
Chengrun Yang, Ziyang Wu, Jerry Chee, Christopher De Sa, Madeleine Udell

Figure 1 for How Low Can We Go: Trading Memory for Error in Low-Precision Training
Figure 2 for How Low Can We Go: Trading Memory for Error in Low-Precision Training
Figure 3 for How Low Can We Go: Trading Memory for Error in Low-Precision Training
Figure 4 for How Low Can We Go: Trading Memory for Error in Low-Precision Training
Viaarxiv icon