Alert button
Picture for Wolfram Burgard

Wolfram Burgard

Alert button

Collaborative Dynamic 3D Scene Graphs for Automated Driving

Sep 19, 2023
Elias Greve, Martin Büchner, Niclas Vödisch, Wolfram Burgard, Abhinav Valada

Figure 1 for Collaborative Dynamic 3D Scene Graphs for Automated Driving
Figure 2 for Collaborative Dynamic 3D Scene Graphs for Automated Driving
Figure 3 for Collaborative Dynamic 3D Scene Graphs for Automated Driving
Figure 4 for Collaborative Dynamic 3D Scene Graphs for Automated Driving

Maps have played an indispensable role in enabling safe and automated driving. Although there have been many advances on different fronts ranging from SLAM to semantics, building an actionable hierarchical semantic representation of urban dynamic scenes from multiple agents is still a challenging problem. In this work, we present Collaborative URBan Scene Graphs (CURB-SG) that enable higher-order reasoning and efficient querying for many functions of automated driving. CURB-SG leverages panoptic LiDAR data from multiple agents to build large-scale maps using an effective graph-based collaborative SLAM approach that detects inter-agent loop closures. To semantically decompose the obtained 3D map, we build a lane graph from the paths of ego agents and their panoptic observations of other vehicles. Based on the connectivity of the lane graph, we segregate the environment into intersecting and non-intersecting road areas. Subsequently, we construct a multi-layered scene graph that includes lane information, the position of static landmarks and their assignment to certain map sections, other vehicles observed by the ego agents, and the pose graph from SLAM including 3D panoptic point clouds. We extensively evaluate CURB-SG in urban scenarios using a photorealistic simulator. We release our code at http://curb.cs.uni-freiburg.de.

* Refined manuscript and extended supplementary 
Viaarxiv icon

Few-Shot Panoptic Segmentation With Foundation Models

Sep 19, 2023
Markus Käppeler, Kürsat Petek, Niclas Vödisch, Wolfram Burgard, Abhinav Valada

Current state-of-the-art methods for panoptic segmentation require an immense amount of annotated training data that is both arduous and expensive to obtain posing a significant challenge for their widespread adoption. Concurrently, recent breakthroughs in visual representation learning have sparked a paradigm shift leading to the advent of large foundation models that can be trained with completely unlabeled images. In this work, we propose to leverage such task-agnostic image features to enable few-shot panoptic segmentation by presenting Segmenting Panoptic Information with Nearly 0 labels (SPINO). In detail, our method combines a DINOv2 backbone with lightweight network heads for semantic segmentation and boundary estimation. We show that our approach, albeit being trained with only ten annotated images, predicts high-quality pseudo-labels that can be used with any existing panoptic segmentation method. Notably, we demonstrate that SPINO achieves competitive results compared to fully supervised baselines while using less than 0.3% of the ground truth labels, paving the way for learning complex visual recognition tasks leveraging foundation models. To illustrate its general applicability, we further deploy SPINO on real-world robotic vision systems for both outdoor and indoor environments. To foster future research, we make the code and trained models publicly available at http://spino.cs.uni-freiburg.de.

Viaarxiv icon

A Smart Robotic System for Industrial Plant Supervision

Sep 01, 2023
D. Adriana Gómez-Rosal, Max Bergau, Georg K. J. Fischer, Andreas Wachaja, Johannes Gräter, Matthias Odenweller, Uwe Piechottka, Fabian Hoeflinger, Nikhil Gosala, Niklas Wetzel, Daniel Büscher, Abhinav Valada, Wolfram Burgard

Figure 1 for A Smart Robotic System for Industrial Plant Supervision
Figure 2 for A Smart Robotic System for Industrial Plant Supervision
Figure 3 for A Smart Robotic System for Industrial Plant Supervision

In today's chemical plants, human field operators perform frequent integrity checks to guarantee high safety standards, and thus are possibly the first to encounter dangerous operating conditions. To alleviate their task, we present a system consisting of an autonomously navigating robot integrated with various sensors and intelligent data processing. It is able to detect methane leaks and estimate its flow rate, detect more general gas anomalies, recognize oil films, localize sound sources and detect failure cases, map the environment in 3D, and navigate autonomously, employing recognition and avoidance of dynamic obstacles. We evaluate our system at a wastewater facility in full working conditions. Our results demonstrate that the system is able to robustly navigate the plant and provide useful information about critical operating conditions.

* Final submission for IEEE Sensors 2023 
Viaarxiv icon

POV-SLAM: Probabilistic Object-Aware Variational SLAM in Semi-Static Environments

Jul 02, 2023
Jingxing Qian, Veronica Chatrath, James Servos, Aaron Mavrinac, Wolfram Burgard, Steven L. Waslander, Angela P. Schoellig

Figure 1 for POV-SLAM: Probabilistic Object-Aware Variational SLAM in Semi-Static Environments
Figure 2 for POV-SLAM: Probabilistic Object-Aware Variational SLAM in Semi-Static Environments
Figure 3 for POV-SLAM: Probabilistic Object-Aware Variational SLAM in Semi-Static Environments
Figure 4 for POV-SLAM: Probabilistic Object-Aware Variational SLAM in Semi-Static Environments

Simultaneous localization and mapping (SLAM) in slowly varying scenes is important for long-term robot task completion. Failing to detect scene changes may lead to inaccurate maps and, ultimately, lost robots. Classical SLAM algorithms assume static scenes, and recent works take dynamics into account, but require scene changes to be observed in consecutive frames. Semi-static scenes, wherein objects appear, disappear, or move slowly over time, are often overlooked, yet are critical for long-term operation. We propose an object-aware, factor-graph SLAM framework that tracks and reconstructs semi-static object-level changes. Our novel variational expectation-maximization strategy is used to optimize factor graphs involving a Gaussian-Uniform bimodal measurement likelihood for potentially-changing objects. We evaluate our approach alongside the state-of-the-art SLAM solutions in simulation and on our novel real-world SLAM dataset captured in a warehouse over four months. Our method improves the robustness of localization in the presence of semi-static changes, providing object-level reasoning about the scene.

* Published in Robotics: Science and Systems (RSS) 2023 
Viaarxiv icon

Geometric Regularity with Robot Intrinsic Symmetry in Reinforcement Learning

Jun 28, 2023
Shengchao Yan, Yuan Zhang, Baohe Zhang, Joschka Boedecker, Wolfram Burgard

Figure 1 for Geometric Regularity with Robot Intrinsic Symmetry in Reinforcement Learning
Figure 2 for Geometric Regularity with Robot Intrinsic Symmetry in Reinforcement Learning
Figure 3 for Geometric Regularity with Robot Intrinsic Symmetry in Reinforcement Learning
Figure 4 for Geometric Regularity with Robot Intrinsic Symmetry in Reinforcement Learning

Geometric regularity, which leverages data symmetry, has been successfully incorporated into deep learning architectures such as CNNs, RNNs, GNNs, and Transformers. While this concept has been widely applied in robotics to address the curse of dimensionality when learning from high-dimensional data, the inherent reflectional and rotational symmetry of robot structures has not been adequately explored. Drawing inspiration from cooperative multi-agent reinforcement learning, we introduce novel network structures for deep learning algorithms that explicitly capture this geometric regularity. Moreover, we investigate the relationship between the geometric prior and the concept of Parameter Sharing in multi-agent reinforcement learning. Through experiments conducted on various challenging continuous control tasks, we demonstrate the significant potential of the proposed geometric regularity in enhancing robot learning capabilities.

* accepted by RSS 2023 Workshop on Symmetries in Robot Learning 
Viaarxiv icon

AutoGraph: Predicting Lane Graphs from Traffic Observations

Jun 27, 2023
Jannik Zürn, Ingmar Posner, Wolfram Burgard

Figure 1 for AutoGraph: Predicting Lane Graphs from Traffic Observations
Figure 2 for AutoGraph: Predicting Lane Graphs from Traffic Observations
Figure 3 for AutoGraph: Predicting Lane Graphs from Traffic Observations
Figure 4 for AutoGraph: Predicting Lane Graphs from Traffic Observations

Lane graph estimation is a long-standing problem in the context of autonomous driving. Previous works aimed at solving this problem by relying on large-scale, hand-annotated lane graphs, introducing a data bottleneck for training models to solve this task. To overcome this limitation, we propose to use the motion patterns of traffic participants as lane graph annotations. In our AutoGraph approach, we employ a pre-trained object tracker to collect the tracklets of traffic participants such as vehicles and trucks. Based on the location of these tracklets, we predict the successor lane graph from an initial position using overhead RGB images only, not requiring any human supervision. In a subsequent stage, we show how the individual successor predictions can be aggregated into a consistent lane graph. We demonstrate the efficacy of our approach on the UrbanLaneGraph dataset and perform extensive quantitative and qualitative evaluations, indicating that AutoGraph is on par with models trained on hand-annotated graph data. Model and dataset will be made available at redacted-for-review.

* 8 pages, 6 figures 
Viaarxiv icon

End-to-end 2D-3D Registration between Image and LiDAR Point Cloud for Vehicle Localization

Jun 20, 2023
Guangming Wang, Yu Zheng, Yanfeng Guo, Zhe Liu, Yixiang Zhu, Wolfram Burgard, Hesheng Wang

Figure 1 for End-to-end 2D-3D Registration between Image and LiDAR Point Cloud for Vehicle Localization
Figure 2 for End-to-end 2D-3D Registration between Image and LiDAR Point Cloud for Vehicle Localization
Figure 3 for End-to-end 2D-3D Registration between Image and LiDAR Point Cloud for Vehicle Localization
Figure 4 for End-to-end 2D-3D Registration between Image and LiDAR Point Cloud for Vehicle Localization

Robot localization using a previously built map is essential for a variety of tasks including highly accurate navigation and mobile manipulation. A popular approach to robot localization is based on image-to-point cloud registration, which combines illumination-invariant LiDAR-based mapping with economical image-based localization. However, the recent works for image-to-point cloud registration either divide the registration into separate modules or project the point cloud to the depth image to register the RGB and depth images. In this paper, we present I2PNet, a novel end-to-end 2D-3D registration network. I2PNet directly registers the raw 3D point cloud with the 2D RGB image using differential modules with a unique target. The 2D-3D cost volume module for differential 2D-3D association is proposed to bridge feature extraction and pose regression. 2D-3D cost volume module implicitly constructs the soft point-to-pixel correspondence on the intrinsic-independent normalized plane of the pinhole camera model. Moreover, we introduce an outlier mask prediction module to filter the outliers in the 2D-3D association before pose regression. Furthermore, we propose the coarse-to-fine 2D-3D registration architecture to increase localization accuracy. We conduct extensive localization experiments on the KITTI Odometry and nuScenes datasets. The results demonstrate that I2PNet outperforms the state-of-the-art by a large margin. In addition, I2PNet has a higher efficiency than the previous works and can perform the localization in real-time. Moreover, we extend the application of I2PNet to the camera-LiDAR online calibration and demonstrate that I2PNet outperforms recent approaches on the online calibration task.

* 18 pages, 14 figures, under review 
Viaarxiv icon

Fast yet predictable braking manoeuvers for real-time robot control

Jun 10, 2023
Mazin Hamad, Jesus Gutierrez-Moreno, Hugo T. M. Kussaba, Nico Mansfeld, Saeed Abdolshah, Abdalla Swikir, Wolfram Burgard, Sami Haddadin

Figure 1 for Fast yet predictable braking manoeuvers for real-time robot control
Figure 2 for Fast yet predictable braking manoeuvers for real-time robot control
Figure 3 for Fast yet predictable braking manoeuvers for real-time robot control
Figure 4 for Fast yet predictable braking manoeuvers for real-time robot control

This paper proposes a framework for generating fast, smooth and predictable braking manoeuvers for a controlled robot. The proposed framework integrates two approaches to obtain feasible modal limits for designing braking trajectories. The first approach is real-time capable but conservative considering the usage of the available feasible actuator control region, resulting in longer braking times. In contrast, the second approach maximizes the used braking control inputs at the cost of requiring more time to evaluate larger, feasible modal limits via optimization. Both approaches allow for predicting the robot's stopping trajectory online. In addition, we also formulated and solved a constrained, nonlinear final-time minimization problem to find optimal torque inputs. The optimal solutions were used as a benchmark to evaluate the performance of the proposed predictable braking framework. A comparative study was compiled in simulation versus a classical optimal controller on a 7-DoF robot arm with only three moving joints. The results verified the effectiveness of our proposed framework and its integrated approaches in achieving fast robot braking manoeuvers with accurate online predictions of the stopping trajectories and distances under various braking settings.

* This work has been accepted to the 22nd IFAC World Congress 
Viaarxiv icon