Picture for David Crandall

David Crandall

SePaint: Semantic Map Inpainting via Multinomial Diffusion

Add code
Mar 05, 2023
Figure 1 for SePaint: Semantic Map Inpainting via Multinomial Diffusion
Figure 2 for SePaint: Semantic Map Inpainting via Multinomial Diffusion
Figure 3 for SePaint: Semantic Map Inpainting via Multinomial Diffusion
Figure 4 for SePaint: Semantic Map Inpainting via Multinomial Diffusion
Viaarxiv icon

LoCoNet: Long-Short Context Network for Active Speaker Detection

Add code
Jan 19, 2023
Figure 1 for LoCoNet: Long-Short Context Network for Active Speaker Detection
Figure 2 for LoCoNet: Long-Short Context Network for Active Speaker Detection
Figure 3 for LoCoNet: Long-Short Context Network for Active Speaker Detection
Figure 4 for LoCoNet: Long-Short Context Network for Active Speaker Detection
Viaarxiv icon

VindLU: A Recipe for Effective Video-and-Language Pretraining

Add code
Dec 09, 2022
Figure 1 for VindLU: A Recipe for Effective Video-and-Language Pretraining
Figure 2 for VindLU: A Recipe for Effective Video-and-Language Pretraining
Figure 3 for VindLU: A Recipe for Effective Video-and-Language Pretraining
Figure 4 for VindLU: A Recipe for Effective Video-and-Language Pretraining
Viaarxiv icon

Attention is All They Need: Exploring the Media Archaeology of the Computer Vision Research Paper

Add code
Sep 22, 2022
Figure 1 for Attention is All They Need: Exploring the Media Archaeology of the Computer Vision Research Paper
Figure 2 for Attention is All They Need: Exploring the Media Archaeology of the Computer Vision Research Paper
Figure 3 for Attention is All They Need: Exploring the Media Archaeology of the Computer Vision Research Paper
Figure 4 for Attention is All They Need: Exploring the Media Archaeology of the Computer Vision Research Paper
Viaarxiv icon

Action Recognition based on Cross-Situational Action-object Statistics

Add code
Aug 15, 2022
Figure 1 for Action Recognition based on Cross-Situational Action-object Statistics
Figure 2 for Action Recognition based on Cross-Situational Action-object Statistics
Figure 3 for Action Recognition based on Cross-Situational Action-object Statistics
Figure 4 for Action Recognition based on Cross-Situational Action-object Statistics
Viaarxiv icon

Graph Neural Network and Spatiotemporal Transformer Attention for 3D Video Object Detection from Point Clouds

Add code
Jul 26, 2022
Figure 1 for Graph Neural Network and Spatiotemporal Transformer Attention for 3D Video Object Detection from Point Clouds
Figure 2 for Graph Neural Network and Spatiotemporal Transformer Attention for 3D Video Object Detection from Point Clouds
Figure 3 for Graph Neural Network and Spatiotemporal Transformer Attention for 3D Video Object Detection from Point Clouds
Figure 4 for Graph Neural Network and Spatiotemporal Transformer Attention for 3D Video Object Detection from Point Clouds
Viaarxiv icon

Reinforcing Generated Images via Meta-learning for One-Shot Fine-Grained Visual Recognition

Add code
Apr 22, 2022
Figure 1 for Reinforcing Generated Images via Meta-learning for One-Shot Fine-Grained Visual Recognition
Figure 2 for Reinforcing Generated Images via Meta-learning for One-Shot Fine-Grained Visual Recognition
Figure 3 for Reinforcing Generated Images via Meta-learning for One-Shot Fine-Grained Visual Recognition
Figure 4 for Reinforcing Generated Images via Meta-learning for One-Shot Fine-Grained Visual Recognition
Viaarxiv icon

Controlling the Quality of Distillation in Response-Based Network Compression

Add code
Dec 19, 2021
Figure 1 for Controlling the Quality of Distillation in Response-Based Network Compression
Figure 2 for Controlling the Quality of Distillation in Response-Based Network Compression
Figure 3 for Controlling the Quality of Distillation in Response-Based Network Compression
Figure 4 for Controlling the Quality of Distillation in Response-Based Network Compression
Viaarxiv icon

Polyline Based Generative Navigable Space Segmentation for Autonomous Visual Navigation

Add code
Oct 29, 2021
Figure 1 for Polyline Based Generative Navigable Space Segmentation for Autonomous Visual Navigation
Figure 2 for Polyline Based Generative Navigable Space Segmentation for Autonomous Visual Navigation
Figure 3 for Polyline Based Generative Navigable Space Segmentation for Autonomous Visual Navigation
Figure 4 for Polyline Based Generative Navigable Space Segmentation for Autonomous Visual Navigation
Viaarxiv icon

Ego4D: Around the World in 3,000 Hours of Egocentric Video

Add code
Oct 13, 2021
Figure 1 for Ego4D: Around the World in 3,000 Hours of Egocentric Video
Figure 2 for Ego4D: Around the World in 3,000 Hours of Egocentric Video
Figure 3 for Ego4D: Around the World in 3,000 Hours of Egocentric Video
Figure 4 for Ego4D: Around the World in 3,000 Hours of Egocentric Video
Viaarxiv icon