Alert button

"Image": models, code, and papers
Alert button

JOTR: 3D Joint Contrastive Learning with Transformers for Occluded Human Mesh Recovery

Add code
Bookmark button
Alert button
Aug 17, 2023
Jiahao Li, Zongxin Yang, Xiaohan Wang, Jianxin Ma, Chang Zhou, Yi Yang

Viaarxiv icon

SurgicalSAM: Efficient Class Promptable Surgical Instrument Segmentation

Add code
Bookmark button
Alert button
Aug 17, 2023
Wenxi Yue, Jing Zhang, Kun Hu, Yong Xia, Jiebo Luo, Zhiyong Wang

Figure 1 for SurgicalSAM: Efficient Class Promptable Surgical Instrument Segmentation
Figure 2 for SurgicalSAM: Efficient Class Promptable Surgical Instrument Segmentation
Figure 3 for SurgicalSAM: Efficient Class Promptable Surgical Instrument Segmentation
Figure 4 for SurgicalSAM: Efficient Class Promptable Surgical Instrument Segmentation
Viaarxiv icon

Text-Only Image Captioning with Multi-Context Data Generation

May 29, 2023
Feipeng Ma, Yizhou Zhou, Fengyun Rao, Yueyi Zhang, Xiaoyan Sun

Figure 1 for Text-Only Image Captioning with Multi-Context Data Generation
Figure 2 for Text-Only Image Captioning with Multi-Context Data Generation
Figure 3 for Text-Only Image Captioning with Multi-Context Data Generation
Figure 4 for Text-Only Image Captioning with Multi-Context Data Generation
Viaarxiv icon

ViCo: Detail-Preserving Visual Condition for Personalized Text-to-Image Generation

Add code
Bookmark button
Alert button
Jun 01, 2023
Shaozhe Hao, Kai Han, Shihao Zhao, Kwan-Yee K. Wong

Figure 1 for ViCo: Detail-Preserving Visual Condition for Personalized Text-to-Image Generation
Figure 2 for ViCo: Detail-Preserving Visual Condition for Personalized Text-to-Image Generation
Figure 3 for ViCo: Detail-Preserving Visual Condition for Personalized Text-to-Image Generation
Figure 4 for ViCo: Detail-Preserving Visual Condition for Personalized Text-to-Image Generation
Viaarxiv icon

Virtual Mirrors: Non-Line-of-Sight Imaging Beyond the Third Bounce

Jul 26, 2023
Diego Royo, Talha Sultan, Adolfo Muñoz, Khadijeh Masumnia-Bisheh, Eric Brandt, Diego Gutierrez, Andreas Velten, Julio Marco

Figure 1 for Virtual Mirrors: Non-Line-of-Sight Imaging Beyond the Third Bounce
Figure 2 for Virtual Mirrors: Non-Line-of-Sight Imaging Beyond the Third Bounce
Figure 3 for Virtual Mirrors: Non-Line-of-Sight Imaging Beyond the Third Bounce
Viaarxiv icon

Optimizing Performance of Feedforward and Convolutional Neural Networks through Dynamic Activation Functions

Aug 10, 2023
Chinmay Rane, Kanishka Tyagi, Michael Manry

Figure 1 for Optimizing Performance of Feedforward and Convolutional Neural Networks through Dynamic Activation Functions
Figure 2 for Optimizing Performance of Feedforward and Convolutional Neural Networks through Dynamic Activation Functions
Figure 3 for Optimizing Performance of Feedforward and Convolutional Neural Networks through Dynamic Activation Functions
Figure 4 for Optimizing Performance of Feedforward and Convolutional Neural Networks through Dynamic Activation Functions
Viaarxiv icon

FuseCap: Leveraging Large Language Models to Fuse Visual Data into Enriched Image Captions

Add code
Bookmark button
Alert button
May 28, 2023
Noam Rotstein, David Bensaid, Shaked Brody, Roy Ganz, Ron Kimmel

Figure 1 for FuseCap: Leveraging Large Language Models to Fuse Visual Data into Enriched Image Captions
Figure 2 for FuseCap: Leveraging Large Language Models to Fuse Visual Data into Enriched Image Captions
Figure 3 for FuseCap: Leveraging Large Language Models to Fuse Visual Data into Enriched Image Captions
Figure 4 for FuseCap: Leveraging Large Language Models to Fuse Visual Data into Enriched Image Captions
Viaarxiv icon

"Let's not Quote out of Context": Unified Vision-Language Pretraining for Context Assisted Image Captioning

Add code
Bookmark button
Alert button
Jun 01, 2023
Abisek Rajakumar Kalarani, Pushpak Bhattacharyya, Niyati Chhaya, Sumit Shekhar

Figure 1 for "Let's not Quote out of Context": Unified Vision-Language Pretraining for Context Assisted Image Captioning
Figure 2 for "Let's not Quote out of Context": Unified Vision-Language Pretraining for Context Assisted Image Captioning
Figure 3 for "Let's not Quote out of Context": Unified Vision-Language Pretraining for Context Assisted Image Captioning
Figure 4 for "Let's not Quote out of Context": Unified Vision-Language Pretraining for Context Assisted Image Captioning
Viaarxiv icon

Study for Performance of MobileNetV1 and MobileNetV2 Based on Breast Cancer

Aug 06, 2023
Jiuqi Yan

Viaarxiv icon

SAMFlow: Eliminating Any Fragmentation in Optical Flow with Segment Anything Model

Aug 16, 2023
Shili Zhou, Ruian He, Weimin Tan, Bo Yan

Figure 1 for SAMFlow: Eliminating Any Fragmentation in Optical Flow with Segment Anything Model
Figure 2 for SAMFlow: Eliminating Any Fragmentation in Optical Flow with Segment Anything Model
Figure 3 for SAMFlow: Eliminating Any Fragmentation in Optical Flow with Segment Anything Model
Figure 4 for SAMFlow: Eliminating Any Fragmentation in Optical Flow with Segment Anything Model
Viaarxiv icon