3d Semantic Segmentation


3D Semantic Segmentation is a computer vision task that involves dividing a 3D point cloud or 3D mesh into semantically meaningful parts or regions. The goal of 3D semantic segmentation is to identify and label different objects and parts within a 3D scene, which can be used for applications such as robotics, autonomous driving, and augmented reality.

Sketch2CT: Multimodal Diffusion for Structure-Aware 3D Medical Volume Generation

Add code
Mar 23, 2026
Viaarxiv icon

FAST3DIS: Feed-forward Anchored Scene Transformer for 3D Instance Segmentation

Add code
Mar 27, 2026
Viaarxiv icon

DriveTok: 3D Driving Scene Tokenization for Unified Multi-View Reconstruction and Understanding

Add code
Mar 19, 2026
Viaarxiv icon

Seeing Where to Deploy: Metric RGB-Based Traversability Analysis for Aerial-to-Ground Hidden Space Inspection

Add code
Mar 15, 2026
Viaarxiv icon

UniSem: Generalizable Semantic 3D Reconstruction from Sparse Unposed Images

Add code
Mar 18, 2026
Viaarxiv icon

Grounding Vision and Language to 3D Masks for Long-Horizon Box Rearrangement

Add code
Mar 24, 2026
Viaarxiv icon

SegFly: A 2D-3D-2D Paradigm for Aerial RGB-Thermal Semantic Segmentation at Scale

Add code
Mar 18, 2026
Viaarxiv icon

Semantic Segmentation and Depth Estimation for Real-Time Lunar Surface Mapping Using 3D Gaussian Splatting

Add code
Mar 18, 2026
Viaarxiv icon

OnlinePG: Online Open-Vocabulary Panoptic Mapping with 3D Gaussian Splatting

Add code
Mar 19, 2026
Viaarxiv icon

Drive-Through 3D Vehicle Exterior Reconstruction via Dynamic-Scene SfM and Distortion-Aware Gaussian Splatting

Add code
Mar 27, 2026
Viaarxiv icon