Picture for Zhen Li

Zhen Li

LMO, CELESTE, HEC Paris

Towards Flexible 3D Perception: Object-Centric Occupancy Completion Augments 3D Object Detection

Add code
Dec 06, 2024
Figure 1 for Towards Flexible 3D Perception: Object-Centric Occupancy Completion Augments 3D Object Detection
Figure 2 for Towards Flexible 3D Perception: Object-Centric Occupancy Completion Augments 3D Object Detection
Figure 3 for Towards Flexible 3D Perception: Object-Centric Occupancy Completion Augments 3D Object Detection
Figure 4 for Towards Flexible 3D Perception: Object-Centric Occupancy Completion Augments 3D Object Detection
Viaarxiv icon

ETSM: Automating Dissection Trajectory Suggestion and Confidence Map-Based Safety Margin Prediction for Robot-assisted Endoscopic Submucosal Dissection

Add code
Nov 28, 2024
Viaarxiv icon

PDZSeg: Adapting the Foundation Model for Dissection Zone Segmentation with Visual Prompts in Robot-assisted Endoscopic Submucosal Dissection

Add code
Nov 27, 2024
Viaarxiv icon

D$^2$-World: An Efficient World Model through Decoupled Dynamic Flow

Add code
Nov 26, 2024
Figure 1 for D$^2$-World: An Efficient World Model through Decoupled Dynamic Flow
Figure 2 for D$^2$-World: An Efficient World Model through Decoupled Dynamic Flow
Figure 3 for D$^2$-World: An Efficient World Model through Decoupled Dynamic Flow
Figure 4 for D$^2$-World: An Efficient World Model through Decoupled Dynamic Flow
Viaarxiv icon

Privacy-Preserving Federated Foundation Model for Generalist Ultrasound Artificial Intelligence

Add code
Nov 25, 2024
Figure 1 for Privacy-Preserving Federated Foundation Model for Generalist Ultrasound Artificial Intelligence
Figure 2 for Privacy-Preserving Federated Foundation Model for Generalist Ultrasound Artificial Intelligence
Figure 3 for Privacy-Preserving Federated Foundation Model for Generalist Ultrasound Artificial Intelligence
Figure 4 for Privacy-Preserving Federated Foundation Model for Generalist Ultrasound Artificial Intelligence
Viaarxiv icon

VisionPAD: A Vision-Centric Pre-training Paradigm for Autonomous Driving

Add code
Nov 22, 2024
Figure 1 for VisionPAD: A Vision-Centric Pre-training Paradigm for Autonomous Driving
Figure 2 for VisionPAD: A Vision-Centric Pre-training Paradigm for Autonomous Driving
Figure 3 for VisionPAD: A Vision-Centric Pre-training Paradigm for Autonomous Driving
Figure 4 for VisionPAD: A Vision-Centric Pre-training Paradigm for Autonomous Driving
Viaarxiv icon

CoPESD: A Multi-Level Surgical Motion Dataset for Training Large Vision-Language Models to Co-Pilot Endoscopic Submucosal Dissection

Add code
Oct 10, 2024
Figure 1 for CoPESD: A Multi-Level Surgical Motion Dataset for Training Large Vision-Language Models to Co-Pilot Endoscopic Submucosal Dissection
Figure 2 for CoPESD: A Multi-Level Surgical Motion Dataset for Training Large Vision-Language Models to Co-Pilot Endoscopic Submucosal Dissection
Figure 3 for CoPESD: A Multi-Level Surgical Motion Dataset for Training Large Vision-Language Models to Co-Pilot Endoscopic Submucosal Dissection
Figure 4 for CoPESD: A Multi-Level Surgical Motion Dataset for Training Large Vision-Language Models to Co-Pilot Endoscopic Submucosal Dissection
Viaarxiv icon

Emu3: Next-Token Prediction is All You Need

Add code
Sep 27, 2024
Figure 1 for Emu3: Next-Token Prediction is All You Need
Figure 2 for Emu3: Next-Token Prediction is All You Need
Figure 3 for Emu3: Next-Token Prediction is All You Need
Figure 4 for Emu3: Next-Token Prediction is All You Need
Viaarxiv icon

MixPolyp: Integrating Mask, Box and Scribble Supervision for Enhanced Polyp Segmentation

Add code
Sep 25, 2024
Figure 1 for MixPolyp: Integrating Mask, Box and Scribble Supervision for Enhanced Polyp Segmentation
Figure 2 for MixPolyp: Integrating Mask, Box and Scribble Supervision for Enhanced Polyp Segmentation
Figure 3 for MixPolyp: Integrating Mask, Box and Scribble Supervision for Enhanced Polyp Segmentation
Figure 4 for MixPolyp: Integrating Mask, Box and Scribble Supervision for Enhanced Polyp Segmentation
Viaarxiv icon

A three-dimensional force estimation method for the cable-driven soft robot based on monocular images

Add code
Sep 12, 2024
Figure 1 for A three-dimensional force estimation method for the cable-driven soft robot based on monocular images
Figure 2 for A three-dimensional force estimation method for the cable-driven soft robot based on monocular images
Figure 3 for A three-dimensional force estimation method for the cable-driven soft robot based on monocular images
Figure 4 for A three-dimensional force estimation method for the cable-driven soft robot based on monocular images
Viaarxiv icon