Jft 300m


On The Relationship between Visual Anomaly-free and Anomalous Representations

Add code
Oct 09, 2024
Viaarxiv icon

Visual Atoms: Pre-training Vision Transformers with Sinusoidal Waves

Add code
Mar 02, 2023
Viaarxiv icon

A simple, efficient and scalable contrastive masked autoencoder for learning visual representations

Add code
Oct 30, 2022
Figure 1 for A simple, efficient and scalable contrastive masked autoencoder for learning visual representations
Figure 2 for A simple, efficient and scalable contrastive masked autoencoder for learning visual representations
Figure 3 for A simple, efficient and scalable contrastive masked autoencoder for learning visual representations
Figure 4 for A simple, efficient and scalable contrastive masked autoencoder for learning visual representations
Viaarxiv icon

Where Should I Spend My FLOPS? Efficiency Evaluations of Visual Pre-training Methods

Add code
Oct 06, 2022
Figure 1 for Where Should I Spend My FLOPS? Efficiency Evaluations of Visual Pre-training Methods
Figure 2 for Where Should I Spend My FLOPS? Efficiency Evaluations of Visual Pre-training Methods
Figure 3 for Where Should I Spend My FLOPS? Efficiency Evaluations of Visual Pre-training Methods
Figure 4 for Where Should I Spend My FLOPS? Efficiency Evaluations of Visual Pre-training Methods
Viaarxiv icon

How to Train Vision Transformer on Small-scale Datasets?

Add code
Oct 13, 2022
Figure 1 for How to Train Vision Transformer on Small-scale Datasets?
Figure 2 for How to Train Vision Transformer on Small-scale Datasets?
Figure 3 for How to Train Vision Transformer on Small-scale Datasets?
Figure 4 for How to Train Vision Transformer on Small-scale Datasets?
Viaarxiv icon

Progress and limitations of deep networks to recognize objects in unusual poses

Add code
Jul 16, 2022
Figure 1 for Progress and limitations of deep networks to recognize objects in unusual poses
Figure 2 for Progress and limitations of deep networks to recognize objects in unusual poses
Figure 3 for Progress and limitations of deep networks to recognize objects in unusual poses
Figure 4 for Progress and limitations of deep networks to recognize objects in unusual poses
Viaarxiv icon

Training Vision Transformers with Only 2040 Images

Add code
Jan 26, 2022
Figure 1 for Training Vision Transformers with Only 2040 Images
Figure 2 for Training Vision Transformers with Only 2040 Images
Figure 3 for Training Vision Transformers with Only 2040 Images
Figure 4 for Training Vision Transformers with Only 2040 Images
Viaarxiv icon

Vision Transformer for Small-Size Datasets

Add code
Dec 27, 2021
Figure 1 for Vision Transformer for Small-Size Datasets
Figure 2 for Vision Transformer for Small-Size Datasets
Figure 3 for Vision Transformer for Small-Size Datasets
Figure 4 for Vision Transformer for Small-Size Datasets
Viaarxiv icon

How to train your ViT? Data, Augmentation, and Regularization in Vision Transformers

Add code
Jun 18, 2021
Figure 1 for How to train your ViT? Data, Augmentation, and Regularization in Vision Transformers
Figure 2 for How to train your ViT? Data, Augmentation, and Regularization in Vision Transformers
Figure 3 for How to train your ViT? Data, Augmentation, and Regularization in Vision Transformers
Figure 4 for How to train your ViT? Data, Augmentation, and Regularization in Vision Transformers
Viaarxiv icon

CoAtNet: Marrying Convolution and Attention for All Data Sizes

Add code
Jun 09, 2021
Figure 1 for CoAtNet: Marrying Convolution and Attention for All Data Sizes
Figure 2 for CoAtNet: Marrying Convolution and Attention for All Data Sizes
Figure 3 for CoAtNet: Marrying Convolution and Attention for All Data Sizes
Figure 4 for CoAtNet: Marrying Convolution and Attention for All Data Sizes
Viaarxiv icon