Picture for Jingjing Li

Jingjing Li

Surgical Scene Segmentation using a Spike-Driven Video Transformer with Real-Time Potential

Add code
Dec 24, 2025
Viaarxiv icon

The Devil is in Attention Sharing: Improving Complex Non-rigid Image Editing Faithfulness via Attention Synergy

Add code
Dec 17, 2025
Viaarxiv icon

Distributed Zero-Shot Learning for Visual Recognition

Add code
Nov 11, 2025
Viaarxiv icon

Acquiring Common Chinese Emotional Events Using Large Language Model

Add code
Nov 07, 2025
Viaarxiv icon

FootFormer: Estimating Stability from Visual Input

Add code
Oct 22, 2025
Viaarxiv icon

ViTs: Teaching Machines to See Time Series Anomalies Like Human Experts

Add code
Oct 06, 2025
Figure 1 for ViTs: Teaching Machines to See Time Series Anomalies Like Human Experts
Figure 2 for ViTs: Teaching Machines to See Time Series Anomalies Like Human Experts
Figure 3 for ViTs: Teaching Machines to See Time Series Anomalies Like Human Experts
Figure 4 for ViTs: Teaching Machines to See Time Series Anomalies Like Human Experts
Viaarxiv icon

Fast, Slow, and Tool-augmented Thinking for LLMs: A Review

Add code
Aug 17, 2025
Viaarxiv icon

Unified modality separation: A vision-language framework for unsupervised domain adaptation

Add code
Aug 07, 2025
Figure 1 for Unified modality separation: A vision-language framework for unsupervised domain adaptation
Figure 2 for Unified modality separation: A vision-language framework for unsupervised domain adaptation
Figure 3 for Unified modality separation: A vision-language framework for unsupervised domain adaptation
Figure 4 for Unified modality separation: A vision-language framework for unsupervised domain adaptation
Viaarxiv icon

SpikeVideoFormer: An Efficient Spike-Driven Video Transformer with Hamming Attention and $\mathcal{O}(T)$ Complexity

Add code
May 15, 2025
Figure 1 for SpikeVideoFormer: An Efficient Spike-Driven Video Transformer with Hamming Attention and $\mathcal{O}(T)$ Complexity
Figure 2 for SpikeVideoFormer: An Efficient Spike-Driven Video Transformer with Hamming Attention and $\mathcal{O}(T)$ Complexity
Figure 3 for SpikeVideoFormer: An Efficient Spike-Driven Video Transformer with Hamming Attention and $\mathcal{O}(T)$ Complexity
Figure 4 for SpikeVideoFormer: An Efficient Spike-Driven Video Transformer with Hamming Attention and $\mathcal{O}(T)$ Complexity
Viaarxiv icon

ReasoningV: Efficient Verilog Code Generation with Adaptive Hybrid Reasoning Model

Add code
Apr 20, 2025
Figure 1 for ReasoningV: Efficient Verilog Code Generation with Adaptive Hybrid Reasoning Model
Figure 2 for ReasoningV: Efficient Verilog Code Generation with Adaptive Hybrid Reasoning Model
Figure 3 for ReasoningV: Efficient Verilog Code Generation with Adaptive Hybrid Reasoning Model
Figure 4 for ReasoningV: Efficient Verilog Code Generation with Adaptive Hybrid Reasoning Model
Viaarxiv icon