Bladder cancer ranks within the top 10 most diagnosed cancers worldwide and is among the most expensive cancers to treat due to the high recurrence rates which require lifetime follow-ups. The primary tool for diagnosis is cystoscopy, which heavily relies on doctors' expertise and interpretation. Therefore, annually, numerous cases are either undiagnosed or misdiagnosed and treated as urinary infections. To address this, we suggest a deep learning approach for bladder cancer detection and segmentation which combines CNNs with a lightweight positional-encoding-free transformer and dual attention gates that fuse self and spatial attention for feature enhancement. The architecture suggested in this paper is efficient making it suitable for medical scenarios that require real time inference. Experiments have proven that this model addresses the critical need for a balance between computational efficiency and diagnostic accuracy in cystoscopic imaging as despite its small size it rivals large models in performance.
The integration of Multi-Rotor Aerial Vehicles (MRAVs) into 5G and 6G networks enhances coverage, connectivity, and congestion management. This fosters communication-aware robotics, exploring the interplay between robotics and communications, but also makes the MRAVs susceptible to malicious attacks, such as jamming. One traditional approach to counter these attacks is the use of beamforming on the MRAVs to apply physical layer security techniques. In this paper, we explore pose optimization as an alternative approach to countering jamming attacks on MRAVs. This technique is intended for omnidirectional MRAVs, which are drones capable of independently controlling both their position and orientation, as opposed to the more common underactuated MRAVs whose orientation cannot be controlled independently of their position. In this paper, we consider an omnidirectional MRAV serving as a Base Station (BS) for legitimate ground nodes, under attack by a malicious jammer. We optimize the MRAV pose (i.e., position and orientation) to maximize the minimum Signal-to-Interference-plus-Noise Ratio (SINR) over all legitimate nodes.
According to the World Health Organization (WHO), air pollution kills seven million people every year. Outdoor air pollution is a major environmental health problem affecting low, middle, and high-income countries. In the past few years, the research community has explored IoT-enabled machine learning applications for outdoor air pollution prediction. The general objective of this paper is to systematically review applications of machine learning and Internet of Things (IoT) for outdoor air pollution prediction and the combination of monitoring sensors and input features used. Two research questions were formulated for this review. 1086 publications were collected in the initial PRISMA stage. After the screening and eligibility phases, 37 papers were selected for inclusion. A cost-based analysis was conducted on the findings to highlight high-cost monitoring, low-cost IoT and hybrid enabled prediction. Three methods of prediction were identified: time series, feature-based and spatio-temporal. This review's findings identify major limitations in applications found in the literature, namely lack of coverage, lack of diversity of data and lack of inclusion of context-specific features. This review proposes directions for future research and underlines practical implications in healthcare, urban planning, global synergy and smart cities.
Internet of Things (IoT) sensors are nowadays heavily utilized in various real-world applications ranging from wearables to smart buildings passing by agrotechnology and health monitoring. With the huge amounts of data generated by these tiny devices, Deep Learning (DL) models have been extensively used to enhance them with intelligent processing. However, with the urge for smaller and more accurate devices, DL models became too heavy to deploy. It is thus necessary to incorporate the hardware's limited resources in the design process. Therefore, inspired by the human brain known for its efficiency and low power consumption, we propose a shallow bidirectional network based on predictive coding theory and dynamic early exiting for halting further computations when a performance threshold is surpassed. We achieve comparable accuracy to VGG-16 in image classification on CIFAR-10 with fewer parameters and less computational complexity.
Although deep learning has made strides in the field of deep noise suppression, leveraging deep architectures on resource-constrained devices still proved challenging. Therefore, we present an early-exiting model based on nsNet2 that provides several levels of accuracy and resource savings by halting computations at different stages. Moreover, we adapt the original architecture by splitting the information flow to take into account the injected dynamism. We show the trade-offs between performance and computational complexity based on established metrics.
Offline Reinforcement Learning (RL) is structured to derive policies from static trajectory data without requiring real-time environment interactions. Recent studies have shown the feasibility of framing offline RL as a sequence modeling task, where the sole aim is to predict actions based on prior context using the transformer architecture. However, the limitation of this single task learning approach is its potential to undermine the transformer model's attention mechanism, which should ideally allocate varying attention weights across different tokens in the input context for optimal prediction. To address this, we reformulate offline RL as a multi-objective optimization problem, where the prediction is extended to states and returns. We also highlight a potential flaw in the trajectory representation used for sequence modeling, which could generate inaccuracies when modeling the state and return distributions. This is due to the non-smoothness of the action distribution within the trajectory dictated by the behavioral policy. To mitigate this issue, we introduce action space regions to the trajectory representation. Our experiments on D4RL benchmark locomotion tasks reveal that our propositions allow for more effective utilization of the attention mechanism in the transformer model, resulting in performance that either matches or outperforms current state-of-the art methods.
The use of Unmanned Ground Vehicles (UGVs) and Unmanned Aerial Vehicles (UAVs) has seen significant growth in the research community, industry, and society. Many of these agents are equipped with communication systems that are essential for completing certain tasks successfully. This has led to the emergence of a new interdisciplinary field at the intersection of robotics and communications, which has been further driven by the integration of UAVs into 5G and 6G communication networks. However, one of the main challenges in this research area is how many researchers tend to oversimplify either the robotics or the communications aspects, hindering the full potential of this new interdisciplinary field. In this paper, we present some of the necessary modeling tools for addressing these problems from both a robotics and communications perspective, using the UAV communications relay as an example.
Fluid mechanics is a fundamental field in engineering and science. Solving the Navier-Stokes equation (NSE) is critical for understanding the behavior of fluids. However, the NSE is a complex partial differential equation that is difficult to solve, and classical numerical methods can be computationally expensive. In this paper, we present an innovative approach for solving the NSE using Physics Informed Neural Networks (PINN) and several novel techniques that improve their performance. The first model is based on an assumption that involves approximating the velocity component by employing the derivative of a stream function. This assumption serves to simplify the system and guarantees that the velocity adheres to the divergence-free equation. We also developed a second more flexible model that approximates the solution without any assumptions. The proposed models can effectively solve two-dimensional NSE. Moreover, we successfully applied the second model to solve the three-dimensional NSE. The results show that the models can efficiently and accurately solve the NSE in three dimensions. These approaches offer several advantages, including high trainability, flexibility, and efficiency.
Mutual relative localization and identification is an important feature for the stabilization and navigation of multi-Unmanned Aerial Vehicle (UAV) systems. Camera-based communications technology, also referred to as Optical Camera Communications (OCC) in the literature, is a novel approach that could bring a valuable solution to such a complex task. In such system, the UAVs are equipped with LEDs that act as beacons and with cameras allowing them to locate the LEDs of other UAVs. Specific blinking sequences are assigned to the LEDs of each of the UAVs in order to uniquely identify them. This camera-based relative localization and identification system is immune to Radio Frequency (RF) electromagnetic interference and operates in Global Navigation satellite (GNSS) denied environments. In addition, since many UAVs are already equipped with cameras, the implementation of this system is inexpensive. In this article, we study in detail the capacity of this system and its limitations. Furthermore, we show how to construct blinking sequences for UAV LEDs in order to improve system performance. Finally, experimental results are presented to corroborate the analytical derivations.
Multi-rotor Unmanned Aerial Vehicles (UAVs) need to tilt in order to move; this modifies the UAV's antenna orientation. We consider the scenario where a multi-rotor UAV serves as a communication relay between a Base Station (BS) and another UAV. We propose a framework to generate feasible trajectories for the multi-rotor UAV relay while considering its motion dynamics and the motion-induced changes of the antenna orientation. The UAV relay's trajectory is optimized to maximize the end-to-end number of bits transmitted. Numerical simulations in MATLAB and Gazebo show the benefits of accounting for the antenna orientation variations due to the UAV tilt.