Picture for Dongchen Han

Dongchen Han

Vision Transformers are Circulant Attention Learners

Add code
Dec 25, 2025
Viaarxiv icon

Step by Step Network

Add code
Nov 18, 2025
Viaarxiv icon

Bridging the Divide: Reconsidering Softmax and Linear Attention

Add code
Dec 09, 2024
Figure 1 for Bridging the Divide: Reconsidering Softmax and Linear Attention
Figure 2 for Bridging the Divide: Reconsidering Softmax and Linear Attention
Figure 3 for Bridging the Divide: Reconsidering Softmax and Linear Attention
Figure 4 for Bridging the Divide: Reconsidering Softmax and Linear Attention
Viaarxiv icon

Efficient Diffusion Transformer with Step-wise Dynamic Attention Mediators

Add code
Aug 11, 2024
Figure 1 for Efficient Diffusion Transformer with Step-wise Dynamic Attention Mediators
Figure 2 for Efficient Diffusion Transformer with Step-wise Dynamic Attention Mediators
Figure 3 for Efficient Diffusion Transformer with Step-wise Dynamic Attention Mediators
Figure 4 for Efficient Diffusion Transformer with Step-wise Dynamic Attention Mediators
Viaarxiv icon

Demystify Mamba in Vision: A Linear Attention Perspective

Add code
May 26, 2024
Figure 1 for Demystify Mamba in Vision: A Linear Attention Perspective
Figure 2 for Demystify Mamba in Vision: A Linear Attention Perspective
Figure 3 for Demystify Mamba in Vision: A Linear Attention Perspective
Figure 4 for Demystify Mamba in Vision: A Linear Attention Perspective
Viaarxiv icon

VL-Trojan: Multimodal Instruction Backdoor Attacks against Autoregressive Visual Language Models

Add code
Feb 21, 2024
Figure 1 for VL-Trojan: Multimodal Instruction Backdoor Attacks against Autoregressive Visual Language Models
Figure 2 for VL-Trojan: Multimodal Instruction Backdoor Attacks against Autoregressive Visual Language Models
Figure 3 for VL-Trojan: Multimodal Instruction Backdoor Attacks against Autoregressive Visual Language Models
Figure 4 for VL-Trojan: Multimodal Instruction Backdoor Attacks against Autoregressive Visual Language Models
Viaarxiv icon

Agent Attention: On the Integration of Softmax and Linear Attention

Add code
Dec 22, 2023
Figure 1 for Agent Attention: On the Integration of Softmax and Linear Attention
Figure 2 for Agent Attention: On the Integration of Softmax and Linear Attention
Figure 3 for Agent Attention: On the Integration of Softmax and Linear Attention
Figure 4 for Agent Attention: On the Integration of Softmax and Linear Attention
Viaarxiv icon

GSVA: Generalized Segmentation via Multimodal Large Language Models

Add code
Dec 15, 2023
Figure 1 for GSVA: Generalized Segmentation via Multimodal Large Language Models
Figure 2 for GSVA: Generalized Segmentation via Multimodal Large Language Models
Figure 3 for GSVA: Generalized Segmentation via Multimodal Large Language Models
Figure 4 for GSVA: Generalized Segmentation via Multimodal Large Language Models
Viaarxiv icon

OT-Attack: Enhancing Adversarial Transferability of Vision-Language Models via Optimal Transport Optimization

Add code
Dec 07, 2023
Figure 1 for OT-Attack: Enhancing Adversarial Transferability of Vision-Language Models via Optimal Transport Optimization
Figure 2 for OT-Attack: Enhancing Adversarial Transferability of Vision-Language Models via Optimal Transport Optimization
Figure 3 for OT-Attack: Enhancing Adversarial Transferability of Vision-Language Models via Optimal Transport Optimization
Figure 4 for OT-Attack: Enhancing Adversarial Transferability of Vision-Language Models via Optimal Transport Optimization
Viaarxiv icon

FLatten Transformer: Vision Transformer using Focused Linear Attention

Add code
Aug 01, 2023
Viaarxiv icon