Alert button
Picture for Klaus Dietmayer

Klaus Dietmayer

Alert button

Simultaneous Clutter Detection and Semantic Segmentation of Moving Objects for Automotive Radar Data

Nov 14, 2023
Johannes Kopp, Dominik Kellner, Aldi Piroli, Vinzenz Dallabetta, Klaus Dietmayer

The unique properties of radar sensors, such as their robustness to adverse weather conditions, make them an important part of the environment perception system of autonomous vehicles. One of the first steps during the processing of radar point clouds is often the detection of clutter, i.e. erroneous points that do not correspond to real objects. Another common objective is the semantic segmentation of moving road users. These two problems are handled strictly separate from each other in literature. The employed neural networks are always focused entirely on only one of the tasks. In contrast to this, we examine ways to solve both tasks at the same time with a single jointly used model. In addition to a new augmented multi-head architecture, we also devise a method to represent a network's predictions for the two tasks with only one output value. This novel approach allows us to solve the tasks simultaneously with the same inference time as a conventional task-specific model. In an extensive evaluation, we show that our setup is highly effective and outperforms every existing network for semantic segmentation on the RadarScenes dataset.

* Published at IEEE International Conference on Intelligent Transportation Systems (ITSC), Bilbao, ESP, 2023 
Viaarxiv icon

Multimodal Object Query Initialization for 3D Object Detection

Oct 16, 2023
Mathijs R. van Geerenstein, Felicia Ruppel, Klaus Dietmayer, Dariu M. Gavrila

3D object detection models that exploit both LiDAR and camera sensor features are top performers in large-scale autonomous driving benchmarks. A transformer is a popular network architecture used for this task, in which so-called object queries act as candidate objects. Initializing these object queries based on current sensor inputs is a common practice. For this, existing methods strongly rely on LiDAR data however, and do not fully exploit image features. Besides, they introduce significant latency. To overcome these limitations we propose EfficientQ3M, an efficient, modular, and multimodal solution for object query initialization for transformer-based 3D object detection models. The proposed initialization method is combined with a "modality-balanced" transformer decoder where the queries can access all sensor modalities throughout the decoder. In experiments, we outperform the state of the art in transformer-based LiDAR object detection on the competitive nuScenes benchmark and showcase the benefits of input-dependent multimodal query initialization, while being more efficient than the available alternatives for LiDAR-camera initialization. The proposed method can be applied with any combination of sensor modalities as input, demonstrating its modularity.

* This work has been submitted to the IEEE for possible publication. Copyright may be transferred without notice, after which this version may no longer be accessible 
Viaarxiv icon

Efficient Path Planning in Large Unknown Environments with Switchable System Models for Automated Vehicles

Oct 10, 2023
Oliver Schumann, Michael Buchholz, Klaus Dietmayer

Figure 1 for Efficient Path Planning in Large Unknown Environments with Switchable System Models for Automated Vehicles
Figure 2 for Efficient Path Planning in Large Unknown Environments with Switchable System Models for Automated Vehicles
Figure 3 for Efficient Path Planning in Large Unknown Environments with Switchable System Models for Automated Vehicles
Figure 4 for Efficient Path Planning in Large Unknown Environments with Switchable System Models for Automated Vehicles

Large environments are challenging for path planning algorithms as the size of the configuration space increases. Furthermore, if the environment is mainly unexplored, large amounts of the path are planned through unknown areas. Hence, a complete replanning of the entire path occurs whenever the path collides with newly discovered obstacles. We propose a novel method that stops the path planning algorithm after a certain distance. It is used to navigate the algorithm in large environments and is not prone to problems of existing navigation approaches. Furthermore, we developed a method to detect significant environment changes to allow a more efficient replanning. At last, we extend the path planner to be used in the U-Shift concept vehicle. It can switch to another system model and rotate around the center of its rear axis. The results show that the proposed methods generate nearly identical paths compared to the standard Hybrid A* while drastically reducing the execution time. Furthermore, we show that the extended path planning algorithm enables the efficient use of the maneuvering capabilities of the concept vehicle to plan concise paths in narrow environments.

Viaarxiv icon

Towards Robust 3D Object Detection In Rainy Conditions

Oct 05, 2023
Aldi Piroli, Vinzenz Dallabetta, Johannes Kopp, Marc Walessa, Daniel Meissner, Klaus Dietmayer

Figure 1 for Towards Robust 3D Object Detection In Rainy Conditions
Figure 2 for Towards Robust 3D Object Detection In Rainy Conditions
Figure 3 for Towards Robust 3D Object Detection In Rainy Conditions
Figure 4 for Towards Robust 3D Object Detection In Rainy Conditions

LiDAR sensors are used in autonomous driving applications to accurately perceive the environment. However, they are affected by adverse weather conditions such as snow, fog, and rain. These everyday phenomena introduce unwanted noise into the measurements, severely degrading the performance of LiDAR-based perception systems. In this work, we propose a framework for improving the robustness of LiDAR-based 3D object detectors against road spray. Our approach uses a state-of-the-art adverse weather detection network to filter out spray from the LiDAR point cloud, which is then used as input for the object detector. In this way, the detected objects are less affected by the adverse weather in the scene, resulting in a more accurate perception of the environment. In addition to adverse weather filtering, we explore the use of radar targets to further filter false positive detections. Tests on real-world data show that our approach improves the robustness to road spray of several popular 3D object detectors.

* Published at IEEE International Conference on Intelligent Transportation Systems ITSC 2023 
Viaarxiv icon

LS-VOS: Identifying Outliers in 3D Object Detections Using Latent Space Virtual Outlier Synthesis

Oct 02, 2023
Aldi Piroli, Vinzenz Dallabetta, Johannes Kopp, Marc Walessa, Daniel Meissner, Klaus Dietmayer

Figure 1 for LS-VOS: Identifying Outliers in 3D Object Detections Using Latent Space Virtual Outlier Synthesis
Figure 2 for LS-VOS: Identifying Outliers in 3D Object Detections Using Latent Space Virtual Outlier Synthesis
Figure 3 for LS-VOS: Identifying Outliers in 3D Object Detections Using Latent Space Virtual Outlier Synthesis
Figure 4 for LS-VOS: Identifying Outliers in 3D Object Detections Using Latent Space Virtual Outlier Synthesis

LiDAR-based 3D object detectors have achieved unprecedented speed and accuracy in autonomous driving applications. However, similar to other neural networks, they are often biased toward high-confidence predictions or return detections where no real object is present. These types of detections can lead to a less reliable environment perception, severely affecting the functionality and safety of autonomous vehicles. We address this problem by proposing LS-VOS, a framework for identifying outliers in 3D object detections. Our approach builds on the idea of Virtual Outlier Synthesis (VOS), which incorporates outlier knowledge during training, enabling the model to learn more compact decision boundaries. In particular, we propose a new synthesis approach that relies on the latent space of an auto-encoder network to generate outlier features with a parametrizable degree of similarity to in-distribution features. In extensive experiments, we show that our approach improves the outlier detection capabilities of a state-of-the-art object detector while maintaining high 3D object detection performance.

* Published at IEEE International Conference on Intelligent Transportation Systems ITSC 2023 
Viaarxiv icon

Group Regression for Query Based Object Detection and Tracking

Aug 28, 2023
Felicia Ruppel, Florian Faion, Claudius Gläser, Klaus Dietmayer

Figure 1 for Group Regression for Query Based Object Detection and Tracking
Figure 2 for Group Regression for Query Based Object Detection and Tracking
Figure 3 for Group Regression for Query Based Object Detection and Tracking
Figure 4 for Group Regression for Query Based Object Detection and Tracking

Group regression is commonly used in 3D object detection to predict box parameters of similar classes in a joint head, aiming to benefit from similarities while separating highly dissimilar classes. For query-based perception methods, this has, so far, not been feasible. We close this gap and present a method to incorporate multi-class group regression, especially designed for the 3D domain in the context of autonomous driving, into existing attention and query-based perception approaches. We enhance a transformer based joint object detection and tracking model with this approach, and thoroughly evaluate its behavior and performance. For group regression, the classes of the nuScenes dataset are divided into six groups of similar shape and prevalence, each being regressed by a dedicated head. We show that the proposed method is applicable to many existing transformer based perception approaches and can bring potential benefits. The behavior of query group regression is thoroughly analyzed in comparison to a unified regression head, e.g. in terms of class-switching behavior and distribution of the output parameters. The proposed method offers many possibilities for further research, such as in the direction of deep multi-hypotheses tracking.

* Accepted for publication at the 2023 26th IEEE International Conference on Intelligent Transportation Systems (ITSC 2023), Sep 24-28, 2023, in Bilbao, Spain 
Viaarxiv icon

User Feedback and Sample Weighting for Ill-Conditioned Hand-Eye Calibration

Aug 11, 2023
Markus Horn, Thomas Wodtko, Michael Buchholz, Klaus Dietmayer

Figure 1 for User Feedback and Sample Weighting for Ill-Conditioned Hand-Eye Calibration
Figure 2 for User Feedback and Sample Weighting for Ill-Conditioned Hand-Eye Calibration
Figure 3 for User Feedback and Sample Weighting for Ill-Conditioned Hand-Eye Calibration
Figure 4 for User Feedback and Sample Weighting for Ill-Conditioned Hand-Eye Calibration

Hand-eye calibration is an important and extensively researched method for calibrating rigidly coupled sensors, solely based on estimates of their motion. Due to the geometric structure of this problem, at least two motion estimates with non-parallel rotation axes are required for a unique solution. If the majority of rotation axes are almost parallel, the resulting optimization problem is ill-conditioned. In this paper, we propose an approach to automatically weight the motion samples of such an ill-conditioned optimization problem for improving the conditioning. The sample weights are chosen in relation to the local density of all available rotation axes. Furthermore, we present an approach for estimating the sensitivity and conditioning of the cost function, separated into the translation and the rotation part. This information can be employed as user feedback when recording the calibration data to prevent ill-conditioning in advance. We evaluate and compare our approach on artificially augmented data from the KITTI odometry dataset.

* Accepted for publication at the 26th IEEE International Conference on Intelligent Transportation Systems 2023 
Viaarxiv icon

Joint Out-of-Distribution Detection and Uncertainty Estimation for Trajectory Prediction

Aug 04, 2023
Julian Wiederer, Julian Schmidt, Ulrich Kressel, Klaus Dietmayer, Vasileios Belagiannis

Figure 1 for Joint Out-of-Distribution Detection and Uncertainty Estimation for Trajectory Prediction
Figure 2 for Joint Out-of-Distribution Detection and Uncertainty Estimation for Trajectory Prediction
Figure 3 for Joint Out-of-Distribution Detection and Uncertainty Estimation for Trajectory Prediction
Figure 4 for Joint Out-of-Distribution Detection and Uncertainty Estimation for Trajectory Prediction

Despite the significant research efforts on trajectory prediction for automated driving, limited work exists on assessing the prediction reliability. To address this limitation we propose an approach that covers two sources of error, namely novel situations with out-of-distribution (OOD) detection and the complexity in in-distribution (ID) situations with uncertainty estimation. We introduce two modules next to an encoder-decoder network for trajectory prediction. Firstly, a Gaussian mixture model learns the probability density function of the ID encoder features during training, and then it is used to detect the OOD samples in regions of the feature space with low likelihood. Secondly, an error regression network is applied to the encoder, which learns to estimate the trajectory prediction error in supervised training. During inference, the estimated prediction error is used as the uncertainty. In our experiments, the combination of both modules outperforms the prior work in OOD detection and uncertainty estimation, on the Shifts robust trajectory prediction dataset by $2.8 \%$ and $10.1 \%$, respectively. The code is publicly available.

* Accepted to the 2023 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2023) 
Viaarxiv icon

Joint Out-of-Distribution Detection and Uncertainty Estimation for Trajectory Predictio

Aug 03, 2023
Julian Wiederer, Julian Schmidt, Ulrich Kressel, Klaus Dietmayer, Vasileios Belagiannis

Figure 1 for Joint Out-of-Distribution Detection and Uncertainty Estimation for Trajectory Predictio
Figure 2 for Joint Out-of-Distribution Detection and Uncertainty Estimation for Trajectory Predictio
Figure 3 for Joint Out-of-Distribution Detection and Uncertainty Estimation for Trajectory Predictio
Figure 4 for Joint Out-of-Distribution Detection and Uncertainty Estimation for Trajectory Predictio

Despite the significant research efforts on trajectory prediction for automated driving, limited work exists on assessing the prediction reliability. To address this limitation we propose an approach that covers two sources of error, namely novel situations with out-of-distribution (OOD) detection and the complexity in in-distribution (ID) situations with uncertainty estimation. We introduce two modules next to an encoder-decoder network for trajectory prediction. Firstly, a Gaussian mixture model learns the probability density function of the ID encoder features during training, and then it is used to detect the OOD samples in regions of the feature space with low likelihood. Secondly, an error regression network is applied to the encoder, which learns to estimate the trajectory prediction error in supervised training. During inference, the estimated prediction error is used as the uncertainty. In our experiments, the combination of both modules outperforms the prior work in OOD detection and uncertainty estimation, on the Shifts robust trajectory prediction dataset by $2.8 \%$ and $10.1 \%$, respectively. The code is publicly available.

* 8 pages, 5 figures, 4 tables (accepted by IROS 2023) 
Viaarxiv icon

Advancing Frame-Dropping in Multi-Object Tracking-by-Detection Systems Through Event-Based Detection Triggering

Aug 01, 2023
Matti Henning, Michael Buchholz, Klaus Dietmayer

Figure 1 for Advancing Frame-Dropping in Multi-Object Tracking-by-Detection Systems Through Event-Based Detection Triggering
Figure 2 for Advancing Frame-Dropping in Multi-Object Tracking-by-Detection Systems Through Event-Based Detection Triggering
Figure 3 for Advancing Frame-Dropping in Multi-Object Tracking-by-Detection Systems Through Event-Based Detection Triggering
Figure 4 for Advancing Frame-Dropping in Multi-Object Tracking-by-Detection Systems Through Event-Based Detection Triggering

With rising computational requirements modern automated vehicles (AVs) often consider trade-offs between energy consumption and perception performance, potentially jeopardizing their safe operation. Frame-dropping in tracking-by-detection perception systems presents a promising approach, although late traffic participant detection might be induced. In this paper, we extend our previous work on frame-dropping in tracking-by-detection perception systems. We introduce an additional event-based triggering mechanism using camera object detections to increase both the system's efficiency, as well as its safety. Evaluating both single and multi-modal tracking methods we show that late object detections are mitigated while the potential for reduced energy consumption is significantly increased, reaching nearly 60 Watt per reduced point in HOTA score.

* Accepted for publication at the 26th IEEE International Conference on Intelligent Transportation Systems 2023 
Viaarxiv icon