Alert button
Picture for Hanxiao Liu

Hanxiao Liu

Alert button

Large Language Models as Optimizers

Sep 07, 2023
Chengrun Yang, Xuezhi Wang, Yifeng Lu, Hanxiao Liu, Quoc V. Le, Denny Zhou, Xinyun Chen

Figure 1 for Large Language Models as Optimizers
Figure 2 for Large Language Models as Optimizers
Figure 3 for Large Language Models as Optimizers
Figure 4 for Large Language Models as Optimizers
Viaarxiv icon

DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining

May 24, 2023
Sang Michael Xie, Hieu Pham, Xuanyi Dong, Nan Du, Hanxiao Liu, Yifeng Lu, Percy Liang, Quoc V. Le, Tengyu Ma, Adams Wei Yu

Figure 1 for DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining
Figure 2 for DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining
Figure 3 for DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining
Figure 4 for DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining
Viaarxiv icon

IDO-VFI: Identifying Dynamics via Optical Flow Guidance for Video Frame Interpolation with Events

May 18, 2023
Chenyang Shi, Hanxiao Liu, Jing Jin, Wenzhuo Li, Yuzhen Li, Boyi Wei, Yibo Zhang

Figure 1 for IDO-VFI: Identifying Dynamics via Optical Flow Guidance for Video Frame Interpolation with Events
Figure 2 for IDO-VFI: Identifying Dynamics via Optical Flow Guidance for Video Frame Interpolation with Events
Figure 3 for IDO-VFI: Identifying Dynamics via Optical Flow Guidance for Video Frame Interpolation with Events
Figure 4 for IDO-VFI: Identifying Dynamics via Optical Flow Guidance for Video Frame Interpolation with Events
Viaarxiv icon

Larger language models do in-context learning differently

Mar 08, 2023
Jerry Wei, Jason Wei, Yi Tay, Dustin Tran, Albert Webson, Yifeng Lu, Xinyun Chen, Hanxiao Liu, Da Huang, Denny Zhou, Tengyu Ma

Figure 1 for Larger language models do in-context learning differently
Figure 2 for Larger language models do in-context learning differently
Figure 3 for Larger language models do in-context learning differently
Figure 4 for Larger language models do in-context learning differently
Viaarxiv icon

Resource-Constrained Neural Architecture Search on Tabular Datasets

Apr 15, 2022
Chengrun Yang, Gabriel Bender, Hanxiao Liu, Pieter-Jan Kindermans, Madeleine Udell, Yifeng Lu, Quoc Le, Da Huang

Figure 1 for Resource-Constrained Neural Architecture Search on Tabular Datasets
Figure 2 for Resource-Constrained Neural Architecture Search on Tabular Datasets
Figure 3 for Resource-Constrained Neural Architecture Search on Tabular Datasets
Figure 4 for Resource-Constrained Neural Architecture Search on Tabular Datasets
Viaarxiv icon

Transformer Quality in Linear Time

Feb 21, 2022
Weizhe Hua, Zihang Dai, Hanxiao Liu, Quoc V. Le

Figure 1 for Transformer Quality in Linear Time
Figure 2 for Transformer Quality in Linear Time
Figure 3 for Transformer Quality in Linear Time
Figure 4 for Transformer Quality in Linear Time
Viaarxiv icon

Mixture-of-Experts with Expert Choice Routing

Feb 18, 2022
Yanqi Zhou, Tao Lei, Hanxiao Liu, Nan Du, Yanping Huang, Vincent Zhao, Andrew Dai, Zhifeng Chen, Quoc Le, James Laudon

Figure 1 for Mixture-of-Experts with Expert Choice Routing
Figure 2 for Mixture-of-Experts with Expert Choice Routing
Figure 3 for Mixture-of-Experts with Expert Choice Routing
Figure 4 for Mixture-of-Experts with Expert Choice Routing
Viaarxiv icon

Combined Scaling for Zero-shot Transfer Learning

Nov 19, 2021
Hieu Pham, Zihang Dai, Golnaz Ghiasi, Hanxiao Liu, Adams Wei Yu, Minh-Thang Luong, Mingxing Tan, Quoc V. Le

Figure 1 for Combined Scaling for Zero-shot Transfer Learning
Figure 2 for Combined Scaling for Zero-shot Transfer Learning
Figure 3 for Combined Scaling for Zero-shot Transfer Learning
Figure 4 for Combined Scaling for Zero-shot Transfer Learning
Viaarxiv icon

Primer: Searching for Efficient Transformers for Language Modeling

Sep 17, 2021
David R. So, Wojciech Mańke, Hanxiao Liu, Zihang Dai, Noam Shazeer, Quoc V. Le

Figure 1 for Primer: Searching for Efficient Transformers for Language Modeling
Figure 2 for Primer: Searching for Efficient Transformers for Language Modeling
Figure 3 for Primer: Searching for Efficient Transformers for Language Modeling
Figure 4 for Primer: Searching for Efficient Transformers for Language Modeling
Viaarxiv icon

CoAtNet: Marrying Convolution and Attention for All Data Sizes

Jun 09, 2021
Zihang Dai, Hanxiao Liu, Quoc V. Le, Mingxing Tan

Figure 1 for CoAtNet: Marrying Convolution and Attention for All Data Sizes
Figure 2 for CoAtNet: Marrying Convolution and Attention for All Data Sizes
Figure 3 for CoAtNet: Marrying Convolution and Attention for All Data Sizes
Figure 4 for CoAtNet: Marrying Convolution and Attention for All Data Sizes
Viaarxiv icon