Picture for Hanxiao Liu

Hanxiao Liu

Large Language Models as Optimizers

Add code
Sep 07, 2023
Figure 1 for Large Language Models as Optimizers
Figure 2 for Large Language Models as Optimizers
Figure 3 for Large Language Models as Optimizers
Figure 4 for Large Language Models as Optimizers
Viaarxiv icon

DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining

May 24, 2023
Figure 1 for DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining
Figure 2 for DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining
Figure 3 for DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining
Figure 4 for DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining
Viaarxiv icon

IDO-VFI: Identifying Dynamics via Optical Flow Guidance for Video Frame Interpolation with Events

Add code
May 18, 2023
Figure 1 for IDO-VFI: Identifying Dynamics via Optical Flow Guidance for Video Frame Interpolation with Events
Figure 2 for IDO-VFI: Identifying Dynamics via Optical Flow Guidance for Video Frame Interpolation with Events
Figure 3 for IDO-VFI: Identifying Dynamics via Optical Flow Guidance for Video Frame Interpolation with Events
Figure 4 for IDO-VFI: Identifying Dynamics via Optical Flow Guidance for Video Frame Interpolation with Events
Viaarxiv icon

Larger language models do in-context learning differently

Mar 08, 2023
Figure 1 for Larger language models do in-context learning differently
Figure 2 for Larger language models do in-context learning differently
Figure 3 for Larger language models do in-context learning differently
Figure 4 for Larger language models do in-context learning differently
Viaarxiv icon

Resource-Constrained Neural Architecture Search on Tabular Datasets

Add code
Apr 15, 2022
Figure 1 for Resource-Constrained Neural Architecture Search on Tabular Datasets
Figure 2 for Resource-Constrained Neural Architecture Search on Tabular Datasets
Figure 3 for Resource-Constrained Neural Architecture Search on Tabular Datasets
Figure 4 for Resource-Constrained Neural Architecture Search on Tabular Datasets
Viaarxiv icon

Transformer Quality in Linear Time

Feb 21, 2022
Figure 1 for Transformer Quality in Linear Time
Figure 2 for Transformer Quality in Linear Time
Figure 3 for Transformer Quality in Linear Time
Figure 4 for Transformer Quality in Linear Time
Viaarxiv icon

Mixture-of-Experts with Expert Choice Routing

Feb 18, 2022
Figure 1 for Mixture-of-Experts with Expert Choice Routing
Figure 2 for Mixture-of-Experts with Expert Choice Routing
Figure 3 for Mixture-of-Experts with Expert Choice Routing
Figure 4 for Mixture-of-Experts with Expert Choice Routing
Viaarxiv icon

Combined Scaling for Zero-shot Transfer Learning

Nov 19, 2021
Figure 1 for Combined Scaling for Zero-shot Transfer Learning
Figure 2 for Combined Scaling for Zero-shot Transfer Learning
Figure 3 for Combined Scaling for Zero-shot Transfer Learning
Figure 4 for Combined Scaling for Zero-shot Transfer Learning
Viaarxiv icon

Primer: Searching for Efficient Transformers for Language Modeling

Add code
Sep 17, 2021
Figure 1 for Primer: Searching for Efficient Transformers for Language Modeling
Figure 2 for Primer: Searching for Efficient Transformers for Language Modeling
Figure 3 for Primer: Searching for Efficient Transformers for Language Modeling
Figure 4 for Primer: Searching for Efficient Transformers for Language Modeling
Viaarxiv icon

CoAtNet: Marrying Convolution and Attention for All Data Sizes

Jun 09, 2021
Figure 1 for CoAtNet: Marrying Convolution and Attention for All Data Sizes
Figure 2 for CoAtNet: Marrying Convolution and Attention for All Data Sizes
Figure 3 for CoAtNet: Marrying Convolution and Attention for All Data Sizes
Figure 4 for CoAtNet: Marrying Convolution and Attention for All Data Sizes
Viaarxiv icon