This paper presents the design concept, modeling and motion planning solution for the aerial robotic chain. This design represents a configurable robotic system of systems, consisting of multi-linked micro aerial vehicles that simultaneously presents the ability to cross narrow sections, morph its shape, ferry significant payloads, offer the potential of distributed sensing and processing, and allow system extendability. We contribute an approach to address the motion planning problem of such a connected robotic system of systems, making full use of its reconfigurable nature, to find collision free paths in a fast manner despite the increased number of degrees of freedom. The presented approach exploits a library of aerial robotic chain configurations, optimized either for cross-section size or sensor coverage, alongside a probabilistic strategy to sample random shape configurations that may be needed to facilitate continued collision-free navigation. Evaluation studies in simulation involve traversal of constrained and obstacle-laden environments, having narrow corridors and cross sections.
Visual Place Recognition (VPR) has seen significant advances at the frontiers of matching performance and computational superiority over the past few years. However, these evaluations are performed for ground-based mobile platforms and cannot be generalized to aerial platforms. The degree of viewpoint variation experienced by aerial robots is complex, with their processing power and on-board memory limited by payload size and battery ratings. Therefore, in this paper, we collect $8$ state-of-the-art VPR techniques that have been previously evaluated for ground-based platforms and compare them on $2$ recently proposed aerial place recognition datasets with three prime focuses: a) Matching performance b) Processing power consumption c) Projected memory requirements. This gives a birds-eye view of the applicability of contemporary VPR research to aerial robotics and lays down the the nature of challenges for aerial-VPR.
This paper proposes a method for tight fusion of visual, depth and inertial data in order to extend robotic capabilities for navigation in GPS-denied, poorly illuminated, and texture-less environments. Visual and depth information are fused at the feature detection and descriptor extraction levels to augment one sensing modality with the other. These multimodal features are then further integrated with inertial sensor cues using an extended Kalman filter to estimate the robot pose, sensor bias terms, and landmark positions simultaneously as part of the filter state. As demonstrated through a set of hand-held and Micro Aerial Vehicle experiments, the proposed algorithm is shown to perform reliably in challenging visually-degraded environments using RGB-D information from a lightweight and low-cost sensor and data from an IMU.
With an ever-widening domain of aerial robotic applications, including many mission critical tasks such as disaster response operations, search and rescue missions and infrastructure inspections taking place in GPS-denied environments, the need for reliable autonomous operation of aerial robots has become crucial. Operating in GPS-denied areas aerial robots rely on a multitude of sensors to localize and navigate. Visible spectrum cameras are the most commonly used sensors due to their low cost and weight. However, in environments that are visually-degraded such as in conditions of poor illumination, low texture, or presence of obscurants including fog, smoke and dust, the reliability of visible light cameras deteriorates significantly. Nevertheless, maintaining reliable robot navigation in such conditions is essential. In contrast to visible light cameras, thermal cameras offer visibility in the infrared spectrum and can be used in a complementary manner with visible spectrum cameras for robot localization and navigation tasks, without paying the significant weight and power penalty typically associated with carrying other sensors. Exploiting this fact, in this work we present a multi-sensor fusion algorithm for reliable odometry estimation in GPS-denied and degraded visual environments. The proposed method utilizes information from both the visible and thermal spectra for landmark selection and prioritizes feature extraction from informative image regions based on a metric over spatial entropy. Furthermore, inertial sensing cues are integrated to improve the robustness of the odometry estimation process. To verify our solution, a set of challenging experiments were conducted inside a) an obscurant filed machine shop-like industrial environment, as well as b) a dark subterranean mine in the presence of heavy airborne dust.
This paper proposes an approach for fusing direct radiometric data from a thermal camera with inertial measurements to extend the robotic capabilities of aerial robots for navigation in GPS-denied and visually degraded environments in the conditions of darkness and in the presence of airborne obscurants such as dust, fog and smoke. An optimization based approach is developed that jointly minimizes the re-projection error of 3D landmarks and inertial measurement errors. The developed solution is extensively verified against both ground-truth in an indoor laboratory setting, as well as inside an underground mine under severely visually degraded conditions.
For robotic inspection tasks in known environments fiducial markers provide a reliable and low-cost solution for robot localization. However, detection of such markers relies on the quality of RGB camera data, which degrades significantly in the presence of visual obscurants such as fog and smoke. The ability to navigate known environments in the presence of obscurants can be critical for inspection tasks especially, in the aftermath of a disaster. Addressing such a scenario, this work proposes a method for the design of fiducial markers to be used with thermal cameras for the pose estimation of aerial robots. Our low cost markers are designed to work in the long wave infrared spectrum, which is not affected by the presence of obscurants, and can be affixed to any object that has measurable temperature difference with respect to its surroundings. Furthermore, the estimated pose from the fiducial markers is fused with inertial measurements in an extended Kalman filter to remove high frequency noise and error present in the fiducial pose estimates. The proposed markers and the pose estimation method are experimentally evaluated in an obscurant filled environment using an aerial robot carrying a thermal camera.
Autonomous robots are commonly tasked with the problem of area exploration and search for certain targets or artifacts of interest to be tracked. Traditionally, the problem formulation considered is that of complete search and thus - ideally - identification of all targets of interest. An important problem however which is not often addressed is that of time-efficient memoryless search under sparse rewards that may be worth visited any number of items. In this paper we specifically address the largely understudied problem of optimizing the "time-of-arrival" or "time-of-detection" to robotically search for sparsely distributed rewards (detect targets of interest) within large-scale environments and subject to memoryless exploration. At the core of the proposed solution is the fact that a search-based L\'evy walk consisting of a constant velocity search following a L\'evy flight path is optimal for searching sparse and randomly distributed target regions in the lack of map memory. A set of results accompany the presentation of the method, demonstrate its properties and justify the purpose of its use towards large-scale area exploration autonomy.
In this research we present a novel algorithm for background subtraction using a moving camera. Our algorithm is based purely on visual information obtained from a camera mounted on an electric bus, operating in downtown Reno which automatically detects moving objects of interest with the view to provide a fully autonomous vehicle. In our approach we exploit the optical flow vectors generated by the motion of the camera while keeping parameter assumptions a minimum. At first, we estimate the Focus of Expansion, which is used to model and simulate 3D points given the intrinsic parameters of the camera, and perform multiple linear regression to estimate the regression equation parameters and implement on the real data set of every frame to identify moving objects. We validated our algorithm using data taken from a common bus route.
Manipulator dynamics, external forces and moments raise issues in stability and efficient control during aerial manipulation. Additionally, multirotor Micro Aerial Vehicles impose stringent limits on payload, actuation and system states. In view of these challenges, this work addressed the design and control of a 3-DoF serial aerial manipulator for contact inspection. A lightweight design with sufficient dexterous workspace for NDT (Non-Destructive Testing) inspection is presented. This operation requires the regulation of normal force on the inspected point. Contact dynamics have been discussed along with a simulation of the closed-loop dynamics during contact. The simulated controller preserves inherent system nonlinearities and uses a passivity approach to ensure the convergence of error to zero. A transition scheme from free-flight to contact was developed along with the hardware and software frameworks for implementation. This paper concludes with important drawbacks and prospects.