Picture for Lucas Beyer

Lucas Beyer

Dima

Getting ViT in Shape: Scaling Laws for Compute-Optimal Model Design

Add code
May 22, 2023
Figure 1 for Getting ViT in Shape: Scaling Laws for Compute-Optimal Model Design
Figure 2 for Getting ViT in Shape: Scaling Laws for Compute-Optimal Model Design
Figure 3 for Getting ViT in Shape: Scaling Laws for Compute-Optimal Model Design
Figure 4 for Getting ViT in Shape: Scaling Laws for Compute-Optimal Model Design
Viaarxiv icon

Sigmoid Loss for Language Image Pre-Training

Add code
Mar 30, 2023
Figure 1 for Sigmoid Loss for Language Image Pre-Training
Figure 2 for Sigmoid Loss for Language Image Pre-Training
Figure 3 for Sigmoid Loss for Language Image Pre-Training
Figure 4 for Sigmoid Loss for Language Image Pre-Training
Viaarxiv icon

A Study of Autoregressive Decoders for Multi-Tasking in Computer Vision

Add code
Mar 30, 2023
Figure 1 for A Study of Autoregressive Decoders for Multi-Tasking in Computer Vision
Figure 2 for A Study of Autoregressive Decoders for Multi-Tasking in Computer Vision
Figure 3 for A Study of Autoregressive Decoders for Multi-Tasking in Computer Vision
Figure 4 for A Study of Autoregressive Decoders for Multi-Tasking in Computer Vision
Viaarxiv icon

Tuning computer vision models with task rewards

Add code
Feb 16, 2023
Figure 1 for Tuning computer vision models with task rewards
Figure 2 for Tuning computer vision models with task rewards
Figure 3 for Tuning computer vision models with task rewards
Figure 4 for Tuning computer vision models with task rewards
Viaarxiv icon

Scaling Vision Transformers to 22 Billion Parameters

Add code
Feb 10, 2023
Figure 1 for Scaling Vision Transformers to 22 Billion Parameters
Figure 2 for Scaling Vision Transformers to 22 Billion Parameters
Figure 3 for Scaling Vision Transformers to 22 Billion Parameters
Figure 4 for Scaling Vision Transformers to 22 Billion Parameters
Viaarxiv icon

FlexiViT: One Model for All Patch Sizes

Add code
Dec 15, 2022
Figure 1 for FlexiViT: One Model for All Patch Sizes
Figure 2 for FlexiViT: One Model for All Patch Sizes
Figure 3 for FlexiViT: One Model for All Patch Sizes
Figure 4 for FlexiViT: One Model for All Patch Sizes
Viaarxiv icon

VeLO: Training Versatile Learned Optimizers by Scaling Up

Add code
Nov 17, 2022
Figure 1 for VeLO: Training Versatile Learned Optimizers by Scaling Up
Figure 2 for VeLO: Training Versatile Learned Optimizers by Scaling Up
Figure 3 for VeLO: Training Versatile Learned Optimizers by Scaling Up
Figure 4 for VeLO: Training Versatile Learned Optimizers by Scaling Up
Viaarxiv icon

PaLI: A Jointly-Scaled Multilingual Language-Image Model

Add code
Sep 16, 2022
Figure 1 for PaLI: A Jointly-Scaled Multilingual Language-Image Model
Figure 2 for PaLI: A Jointly-Scaled Multilingual Language-Image Model
Figure 3 for PaLI: A Jointly-Scaled Multilingual Language-Image Model
Figure 4 for PaLI: A Jointly-Scaled Multilingual Language-Image Model
Viaarxiv icon

UViM: A Unified Modeling Approach for Vision with Learned Guiding Codes

Add code
May 27, 2022
Figure 1 for UViM: A Unified Modeling Approach for Vision with Learned Guiding Codes
Figure 2 for UViM: A Unified Modeling Approach for Vision with Learned Guiding Codes
Figure 3 for UViM: A Unified Modeling Approach for Vision with Learned Guiding Codes
Figure 4 for UViM: A Unified Modeling Approach for Vision with Learned Guiding Codes
Viaarxiv icon

Better plain ViT baselines for ImageNet-1k

Add code
May 03, 2022
Figure 1 for Better plain ViT baselines for ImageNet-1k
Figure 2 for Better plain ViT baselines for ImageNet-1k
Figure 3 for Better plain ViT baselines for ImageNet-1k
Viaarxiv icon