Picture for Christoph Feichtenhofer

Christoph Feichtenhofer

Window Attention is Bugged: How not to Interpolate Position Embeddings

Add code
Nov 09, 2023
Figure 1 for Window Attention is Bugged: How not to Interpolate Position Embeddings
Figure 2 for Window Attention is Bugged: How not to Interpolate Position Embeddings
Figure 3 for Window Attention is Bugged: How not to Interpolate Position Embeddings
Figure 4 for Window Attention is Bugged: How not to Interpolate Position Embeddings
Viaarxiv icon

Demystifying CLIP Data

Add code
Oct 02, 2023
Figure 1 for Demystifying CLIP Data
Figure 2 for Demystifying CLIP Data
Figure 3 for Demystifying CLIP Data
Figure 4 for Demystifying CLIP Data
Viaarxiv icon

Hiera: A Hierarchical Vision Transformer without the Bells-and-Whistles

Add code
Jun 01, 2023
Figure 1 for Hiera: A Hierarchical Vision Transformer without the Bells-and-Whistles
Figure 2 for Hiera: A Hierarchical Vision Transformer without the Bells-and-Whistles
Figure 3 for Hiera: A Hierarchical Vision Transformer without the Bells-and-Whistles
Figure 4 for Hiera: A Hierarchical Vision Transformer without the Bells-and-Whistles
Viaarxiv icon

Diffusion Models as Masked Autoencoders

Add code
Apr 06, 2023
Figure 1 for Diffusion Models as Masked Autoencoders
Figure 2 for Diffusion Models as Masked Autoencoders
Figure 3 for Diffusion Models as Masked Autoencoders
Figure 4 for Diffusion Models as Masked Autoencoders
Viaarxiv icon

On the Benefits of 3D Pose and Tracking for Human Action Recognition

Add code
Apr 03, 2023
Figure 1 for On the Benefits of 3D Pose and Tracking for Human Action Recognition
Figure 2 for On the Benefits of 3D Pose and Tracking for Human Action Recognition
Figure 3 for On the Benefits of 3D Pose and Tracking for Human Action Recognition
Figure 4 for On the Benefits of 3D Pose and Tracking for Human Action Recognition
Viaarxiv icon

The effectiveness of MAE pre-pretraining for billion-scale pretraining

Add code
Mar 23, 2023
Figure 1 for The effectiveness of MAE pre-pretraining for billion-scale pretraining
Figure 2 for The effectiveness of MAE pre-pretraining for billion-scale pretraining
Figure 3 for The effectiveness of MAE pre-pretraining for billion-scale pretraining
Figure 4 for The effectiveness of MAE pre-pretraining for billion-scale pretraining
Viaarxiv icon

Reversible Vision Transformers

Add code
Feb 09, 2023
Figure 1 for Reversible Vision Transformers
Figure 2 for Reversible Vision Transformers
Figure 3 for Reversible Vision Transformers
Figure 4 for Reversible Vision Transformers
Viaarxiv icon

Multiview Compressive Coding for 3D Reconstruction

Add code
Jan 19, 2023
Figure 1 for Multiview Compressive Coding for 3D Reconstruction
Figure 2 for Multiview Compressive Coding for 3D Reconstruction
Figure 3 for Multiview Compressive Coding for 3D Reconstruction
Figure 4 for Multiview Compressive Coding for 3D Reconstruction
Viaarxiv icon

CiT: Curation in Training for Effective Vision-Language Data

Add code
Jan 05, 2023
Figure 1 for CiT: Curation in Training for Effective Vision-Language Data
Figure 2 for CiT: Curation in Training for Effective Vision-Language Data
Figure 3 for CiT: Curation in Training for Effective Vision-Language Data
Figure 4 for CiT: Curation in Training for Effective Vision-Language Data
Viaarxiv icon

MAViL: Masked Audio-Video Learners

Add code
Dec 15, 2022
Figure 1 for MAViL: Masked Audio-Video Learners
Figure 2 for MAViL: Masked Audio-Video Learners
Figure 3 for MAViL: Masked Audio-Video Learners
Figure 4 for MAViL: Masked Audio-Video Learners
Viaarxiv icon