Picture for Aykut Erdem

Aykut Erdem

Shammie

Sequential Compositional Generalization in Multimodal Models

Add code
Apr 18, 2024
Figure 1 for Sequential Compositional Generalization in Multimodal Models
Figure 2 for Sequential Compositional Generalization in Multimodal Models
Figure 3 for Sequential Compositional Generalization in Multimodal Models
Figure 4 for Sequential Compositional Generalization in Multimodal Models
Viaarxiv icon

ViLMA: A Zero-Shot Benchmark for Linguistic and Temporal Grounding in Video-Language Models

Add code
Nov 13, 2023
Figure 1 for ViLMA: A Zero-Shot Benchmark for Linguistic and Temporal Grounding in Video-Language Models
Figure 2 for ViLMA: A Zero-Shot Benchmark for Linguistic and Temporal Grounding in Video-Language Models
Figure 3 for ViLMA: A Zero-Shot Benchmark for Linguistic and Temporal Grounding in Video-Language Models
Figure 4 for ViLMA: A Zero-Shot Benchmark for Linguistic and Temporal Grounding in Video-Language Models
Viaarxiv icon

Harnessing Dataset Cartography for Improved Compositional Generalization in Transformers

Add code
Oct 18, 2023
Figure 1 for Harnessing Dataset Cartography for Improved Compositional Generalization in Transformers
Figure 2 for Harnessing Dataset Cartography for Improved Compositional Generalization in Transformers
Figure 3 for Harnessing Dataset Cartography for Improved Compositional Generalization in Transformers
Figure 4 for Harnessing Dataset Cartography for Improved Compositional Generalization in Transformers
Viaarxiv icon

Hyperspectral Image Denoising via Self-Modulating Convolutional Neural Networks

Add code
Sep 15, 2023
Figure 1 for Hyperspectral Image Denoising via Self-Modulating Convolutional Neural Networks
Figure 2 for Hyperspectral Image Denoising via Self-Modulating Convolutional Neural Networks
Figure 3 for Hyperspectral Image Denoising via Self-Modulating Convolutional Neural Networks
Figure 4 for Hyperspectral Image Denoising via Self-Modulating Convolutional Neural Networks
Viaarxiv icon

Spherical Vision Transformer for 360-degree Video Saliency Prediction

Add code
Aug 24, 2023
Figure 1 for Spherical Vision Transformer for 360-degree Video Saliency Prediction
Figure 2 for Spherical Vision Transformer for 360-degree Video Saliency Prediction
Figure 3 for Spherical Vision Transformer for 360-degree Video Saliency Prediction
Figure 4 for Spherical Vision Transformer for 360-degree Video Saliency Prediction
Viaarxiv icon

CLIP-Guided StyleGAN Inversion for Text-Driven Real Image Editing

Add code
Jul 18, 2023
Figure 1 for CLIP-Guided StyleGAN Inversion for Text-Driven Real Image Editing
Figure 2 for CLIP-Guided StyleGAN Inversion for Text-Driven Real Image Editing
Figure 3 for CLIP-Guided StyleGAN Inversion for Text-Driven Real Image Editing
Figure 4 for CLIP-Guided StyleGAN Inversion for Text-Driven Real Image Editing
Viaarxiv icon

HyperE2VID: Improving Event-Based Video Reconstruction via Hypernetworks

Add code
May 10, 2023
Viaarxiv icon

EVREAL: Towards a Comprehensive Benchmark and Analysis Suite for Event-based Video Reconstruction

Add code
Apr 30, 2023
Viaarxiv icon

VidStyleODE: Disentangled Video Editing via StyleGAN and NeuralODEs

Add code
Apr 12, 2023
Figure 1 for VidStyleODE: Disentangled Video Editing via StyleGAN and NeuralODEs
Figure 2 for VidStyleODE: Disentangled Video Editing via StyleGAN and NeuralODEs
Figure 3 for VidStyleODE: Disentangled Video Editing via StyleGAN and NeuralODEs
Figure 4 for VidStyleODE: Disentangled Video Editing via StyleGAN and NeuralODEs
Viaarxiv icon

Inst-Inpaint: Instructing to Remove Objects with Diffusion Models

Add code
Apr 06, 2023
Figure 1 for Inst-Inpaint: Instructing to Remove Objects with Diffusion Models
Figure 2 for Inst-Inpaint: Instructing to Remove Objects with Diffusion Models
Figure 3 for Inst-Inpaint: Instructing to Remove Objects with Diffusion Models
Figure 4 for Inst-Inpaint: Instructing to Remove Objects with Diffusion Models
Viaarxiv icon