Picture for Siyuan Qiao

Siyuan Qiao

IG Captioner: Information Gain Captioners are Strong Zero-shot Classifiers

Add code
Nov 27, 2023
Figure 1 for IG Captioner: Information Gain Captioners are Strong Zero-shot Classifiers
Figure 2 for IG Captioner: Information Gain Captioners are Strong Zero-shot Classifiers
Figure 3 for IG Captioner: Information Gain Captioners are Strong Zero-shot Classifiers
Figure 4 for IG Captioner: Information Gain Captioners are Strong Zero-shot Classifiers
Viaarxiv icon

PolyMaX: General Dense Prediction with Mask Transformer

Add code
Nov 09, 2023
Figure 1 for PolyMaX: General Dense Prediction with Mask Transformer
Figure 2 for PolyMaX: General Dense Prediction with Mask Transformer
Figure 3 for PolyMaX: General Dense Prediction with Mask Transformer
Figure 4 for PolyMaX: General Dense Prediction with Mask Transformer
Viaarxiv icon

De-Diffusion Makes Text a Strong Cross-Modal Interface

Add code
Nov 01, 2023
Figure 1 for De-Diffusion Makes Text a Strong Cross-Modal Interface
Figure 2 for De-Diffusion Makes Text a Strong Cross-Modal Interface
Figure 3 for De-Diffusion Makes Text a Strong Cross-Modal Interface
Figure 4 for De-Diffusion Makes Text a Strong Cross-Modal Interface
Viaarxiv icon

Superpixel Transformers for Efficient Semantic Segmentation

Add code
Oct 02, 2023
Figure 1 for Superpixel Transformers for Efficient Semantic Segmentation
Figure 2 for Superpixel Transformers for Efficient Semantic Segmentation
Figure 3 for Superpixel Transformers for Efficient Semantic Segmentation
Figure 4 for Superpixel Transformers for Efficient Semantic Segmentation
Viaarxiv icon

PaLM 2 Technical Report

Add code
May 17, 2023
Figure 1 for PaLM 2 Technical Report
Figure 2 for PaLM 2 Technical Report
Figure 3 for PaLM 2 Technical Report
Figure 4 for PaLM 2 Technical Report
Viaarxiv icon

MOAT: Alternating Mobile Convolution and Attention Brings Strong Vision Models

Add code
Oct 04, 2022
Figure 1 for MOAT: Alternating Mobile Convolution and Attention Brings Strong Vision Models
Figure 2 for MOAT: Alternating Mobile Convolution and Attention Brings Strong Vision Models
Figure 3 for MOAT: Alternating Mobile Convolution and Attention Brings Strong Vision Models
Figure 4 for MOAT: Alternating Mobile Convolution and Attention Brings Strong Vision Models
Viaarxiv icon

k-means Mask Transformer

Add code
Jul 08, 2022
Figure 1 for k-means Mask Transformer
Figure 2 for k-means Mask Transformer
Figure 3 for k-means Mask Transformer
Figure 4 for k-means Mask Transformer
Viaarxiv icon

CMT-DeepLab: Clustering Mask Transformers for Panoptic Segmentation

Add code
Jun 17, 2022
Figure 1 for CMT-DeepLab: Clustering Mask Transformers for Panoptic Segmentation
Figure 2 for CMT-DeepLab: Clustering Mask Transformers for Panoptic Segmentation
Figure 3 for CMT-DeepLab: Clustering Mask Transformers for Panoptic Segmentation
Figure 4 for CMT-DeepLab: Clustering Mask Transformers for Panoptic Segmentation
Viaarxiv icon

Waymo Open Dataset: Panoramic Video Panoptic Segmentation

Add code
Jun 15, 2022
Figure 1 for Waymo Open Dataset: Panoramic Video Panoptic Segmentation
Figure 2 for Waymo Open Dataset: Panoramic Video Panoptic Segmentation
Figure 3 for Waymo Open Dataset: Panoramic Video Panoptic Segmentation
Figure 4 for Waymo Open Dataset: Panoramic Video Panoptic Segmentation
Viaarxiv icon

TubeFormer-DeepLab: Video Mask Transformer

Add code
May 30, 2022
Figure 1 for TubeFormer-DeepLab: Video Mask Transformer
Figure 2 for TubeFormer-DeepLab: Video Mask Transformer
Figure 3 for TubeFormer-DeepLab: Video Mask Transformer
Figure 4 for TubeFormer-DeepLab: Video Mask Transformer
Viaarxiv icon