Alert button

"Image": models, code, and papers
Alert button

Monocular Robot Navigation with Self-Supervised Pretrained Vision Transformers

Add code
Bookmark button
Alert button
Mar 07, 2022
Miguel Saavedra-Ruiz, Sacha Morin, Liam Paull

Figure 1 for Monocular Robot Navigation with Self-Supervised Pretrained Vision Transformers
Figure 2 for Monocular Robot Navigation with Self-Supervised Pretrained Vision Transformers
Figure 3 for Monocular Robot Navigation with Self-Supervised Pretrained Vision Transformers
Figure 4 for Monocular Robot Navigation with Self-Supervised Pretrained Vision Transformers
Viaarxiv icon

UViM: A Unified Modeling Approach for Vision with Learned Guiding Codes

Add code
Bookmark button
Alert button
May 20, 2022
Alexander Kolesnikov, André Susano Pinto, Lucas Beyer, Xiaohua Zhai, Jeremiah Harmsen, Neil Houlsby

Figure 1 for UViM: A Unified Modeling Approach for Vision with Learned Guiding Codes
Figure 2 for UViM: A Unified Modeling Approach for Vision with Learned Guiding Codes
Figure 3 for UViM: A Unified Modeling Approach for Vision with Learned Guiding Codes
Figure 4 for UViM: A Unified Modeling Approach for Vision with Learned Guiding Codes
Viaarxiv icon

Unraveling Attention via Convex Duality: Analysis and Interpretations of Vision Transformers

Add code
Bookmark button
Alert button
May 20, 2022
Arda Sahiner, Tolga Ergen, Batu Ozturkler, John Pauly, Morteza Mardani, Mert Pilanci

Figure 1 for Unraveling Attention via Convex Duality: Analysis and Interpretations of Vision Transformers
Figure 2 for Unraveling Attention via Convex Duality: Analysis and Interpretations of Vision Transformers
Figure 3 for Unraveling Attention via Convex Duality: Analysis and Interpretations of Vision Transformers
Figure 4 for Unraveling Attention via Convex Duality: Analysis and Interpretations of Vision Transformers
Viaarxiv icon

Why do CNNs Learn Consistent Representations in their First Layer Independent of Labels and Architecture?

Jun 06, 2022
Rhea Chowers, Yair Weiss

Figure 1 for Why do CNNs Learn Consistent Representations in their First Layer Independent of Labels and Architecture?
Figure 2 for Why do CNNs Learn Consistent Representations in their First Layer Independent of Labels and Architecture?
Figure 3 for Why do CNNs Learn Consistent Representations in their First Layer Independent of Labels and Architecture?
Figure 4 for Why do CNNs Learn Consistent Representations in their First Layer Independent of Labels and Architecture?
Viaarxiv icon

Convolutional Neural Network to Restore Low-Dose Digital Breast Tomosynthesis Projections in a Variance Stabilization Domain

Mar 22, 2022
Rodrigo de Barros Vimieiro, Chuang Niu, Hongming Shan, Lucas Rodrigues Borges, Ge Wang, Marcelo Andrade da Costa Vieira

Figure 1 for Convolutional Neural Network to Restore Low-Dose Digital Breast Tomosynthesis Projections in a Variance Stabilization Domain
Figure 2 for Convolutional Neural Network to Restore Low-Dose Digital Breast Tomosynthesis Projections in a Variance Stabilization Domain
Figure 3 for Convolutional Neural Network to Restore Low-Dose Digital Breast Tomosynthesis Projections in a Variance Stabilization Domain
Figure 4 for Convolutional Neural Network to Restore Low-Dose Digital Breast Tomosynthesis Projections in a Variance Stabilization Domain
Viaarxiv icon

Instance-level Image Retrieval using Reranking Transformers

Add code
Bookmark button
Alert button
Mar 22, 2021
Fuwen Tan, Jiangbo Yuan, Vicente Ordonez

Figure 1 for Instance-level Image Retrieval using Reranking Transformers
Figure 2 for Instance-level Image Retrieval using Reranking Transformers
Figure 3 for Instance-level Image Retrieval using Reranking Transformers
Figure 4 for Instance-level Image Retrieval using Reranking Transformers
Viaarxiv icon

Visually-Augmented Language Modeling

Add code
Bookmark button
Alert button
May 20, 2022
Weizhi Wang, Li Dong, Hao Cheng, Haoyu Song, Xiaodong Liu, Xifeng Yan, Jianfeng Gao, Furu Wei

Figure 1 for Visually-Augmented Language Modeling
Figure 2 for Visually-Augmented Language Modeling
Figure 3 for Visually-Augmented Language Modeling
Figure 4 for Visually-Augmented Language Modeling
Viaarxiv icon

Where is the disease? Semi-supervised pseudo-normality synthesis from an abnormal image

Jun 24, 2021
Yuanqi Du, Quan Quan, Hu Han, S. Kevin Zhou

Figure 1 for Where is the disease? Semi-supervised pseudo-normality synthesis from an abnormal image
Figure 2 for Where is the disease? Semi-supervised pseudo-normality synthesis from an abnormal image
Figure 3 for Where is the disease? Semi-supervised pseudo-normality synthesis from an abnormal image
Figure 4 for Where is the disease? Semi-supervised pseudo-normality synthesis from an abnormal image
Viaarxiv icon

Glance to Count: Learning to Rank with Anchors for Weakly-supervised Crowd Counting

May 29, 2022
Zheng Xiong, Liangyu Chai, Wenxi Liu, Yongtuo Liu, Sucheng Ren, Shengfeng He

Figure 1 for Glance to Count: Learning to Rank with Anchors for Weakly-supervised Crowd Counting
Figure 2 for Glance to Count: Learning to Rank with Anchors for Weakly-supervised Crowd Counting
Figure 3 for Glance to Count: Learning to Rank with Anchors for Weakly-supervised Crowd Counting
Figure 4 for Glance to Count: Learning to Rank with Anchors for Weakly-supervised Crowd Counting
Viaarxiv icon

Improving Monocular Visual Odometry Using Learned Depth

Apr 04, 2022
Libo Sun, Wei Yin, Enze Xie, Zhengrong Li, Changming Sun, Chunhua Shen

Figure 1 for Improving Monocular Visual Odometry Using Learned Depth
Figure 2 for Improving Monocular Visual Odometry Using Learned Depth
Figure 3 for Improving Monocular Visual Odometry Using Learned Depth
Figure 4 for Improving Monocular Visual Odometry Using Learned Depth
Viaarxiv icon