Spiking neural networks (SNNs) implemented on neuromorphic processors (NPs) can enhance the energy efficiency of deployments of artificial intelligence (AI) for specific workloads. As such, NP represents an interesting opportunity for implementing AI tasks on board power-limited satellite communication spacecraft. In this article, we disseminate the findings of a recently completed study which targeted the comparison in terms of performance and power-consumption of different satellite communication use cases implemented on standard AI accelerators and on NPs. In particular, the article describes three prominent use cases, namely payload resource optimization, onboard interference detection and classification, and dynamic receive beamforming; and compare the performance of conventional convolutional neural networks (CNNs) implemented on Xilinx's VCK5000 Versal development card and SNNs on Intel's neuromorphic chip Loihi 2.
Satellite communications (SatCom) are crucial for global connectivity, especially in the era of emerging technologies like 6G and narrowing the digital divide. Traditional SatCom systems struggle with efficient resource management due to static multibeam configurations, hindering quality of service (QoS) amidst dynamic traffic demands. This paper introduces an innovative solution - real-time adaptive beamforming on multibeam satellites with software-defined payloads in geostationary orbit (GEO). Utilizing a Direct Radiating Array (DRA) with circular polarization in the 17.7 - 20.2 GHz band, the paper outlines DRA design and a supervised learning-based algorithm for on-board beamforming. This adaptive approach not only meets precise beam projection needs but also dynamically adjusts beamwidth, minimizes sidelobe levels (SLL), and optimizes effective isotropic radiated power (EIRP).
Satellite communications, essential for modern connectivity, extend access to maritime, aeronautical, and remote areas where terrestrial networks are unfeasible. Current GEO systems distribute power and bandwidth uniformly across beams using multi-beam footprints with fractional frequency reuse. However, recent research reveals the limitations of this approach in heterogeneous traffic scenarios, leading to inefficiencies. To address this, this paper presents a machine learning (ML)-based approach to Radio Resource Management (RRM). We treat the RRM task as a regression ML problem, integrating RRM objectives and constraints into the loss function that the ML algorithm aims at minimizing. Moreover, we introduce a context-aware ML metric that evaluates the ML model's performance but also considers the impact of its resource allocation decisions on the overall performance of the communication system.
Backscatter communication (BC) technology offers sustainable solutions for next-generation Internet-of-Things (IoT) networks, where devices can transmit data by reflecting and adjusting incident radio frequency signals. In parallel to BC, deep reinforcement learning (DRL) has recently emerged as a promising tool to augment intelligence and optimize low-powered IoT devices. This article commences by elucidating the foundational principles underpinning BC systems, subsequently delving into the diverse array of DRL techniques and their respective practical implementations. Subsequently, it investigates potential domains and presents recent advancements in the realm of DRL-BC systems. A use case of RIS-aided non-orthogonal multiple access BC systems leveraging DRL is meticulously examined to highlight its potential. Lastly, this study identifies and investigates salient challenges and proffers prospective avenues for future research endeavors.
This paper studies the potential of RIS-integrated NTNs to revolutionize the next-generation connectivity. First, it discusses the fundamentals of RIS technology. Secondly, it delves into reporting the recent advances in RIS-enabled NTNs. Subsequently, it presents a novel framework based on the current state-of-the-art for low earth orbit satellites (LEO) communications, wherein the signal received at the user terminal traverses both a direct link and an RIS link, and the RIS is mounted on a high-altitude platform (HAP) situated within the stratosphere. Finally, the paper concludes by highlighting open challenges and future research directions to revolutionize the realm of RIS-integrated NTNs.
The latest satellite communication (SatCom) missions are characterized by a fully reconfigurable on-board software-defined payload, capable of adapting radio resources to the temporal and spatial variations of the system traffic. As pure optimization-based solutions have shown to be computationally tedious and to lack flexibility, machine learning (ML)-based methods have emerged as promising alternatives. We investigate the application of energy-efficient brain-inspired ML models for on-board radio resource management. Apart from software simulation, we report extensive experimental results leveraging the recently released Intel Loihi 2 chip. To benchmark the performance of the proposed model, we implement conventional convolutional neural networks (CNN) on a Xilinx Versal VCK5000, and provide a detailed comparison of accuracy, precision, recall, and energy efficiency for different traffic demands. Most notably, for relevant workloads, spiking neural networks (SNNs) implemented on Loihi 2 yield higher accuracy, while reducing power consumption by more than 100$\times$ as compared to the CNN-based reference platform. Our findings point to the significant potential of neuromorphic computing and SNNs in supporting on-board SatCom operations, paving the way for enhanced efficiency and sustainability in future SatCom systems.
The paradigm of joint communications and sensing (JCAS) envisions a revolutionary integration of communication and radar functionalities within a unified hardware platform. This novel concept not only opens up unprecedented possibilities, but also presents unique challenges. Its success is highly dependent on efficient full-duplex (FD) operation, which has the potential to enable simultaneous transmission and reception within the same frequency band. While ongoing research explores the potential of JCAS, there are related avenues of investigation that hold tremendous potential to profoundly transform the sixth generation (6G) and beyond cellular networks. This article sheds light on the new opportunities and challenges presented by JCAS by taking into account the key technical challenges of FD systems. Unlike simplified JCAS scenarios, we delve into the most comprehensive configuration, encompassing uplink (UL) and downlink (DL) users, as well as monostatic and bistatic radars, all harmoniously coexisting to jointly push the boundaries of both the communications and sensing performance. The performance improvements introduced by this advancement bring forth numerous new challenges, each meticulously examined and expounded upon.
Device-to-device (D2D) communications offers high spectral efficiency, low energy consumption and transmission latency. However, one of the main limitations of D2D communications is co-channel interference from underlaying wireless system. Reconfigurable intelligent surfaces (RIS) is a promising technology because it can manipulate the electromagnetic waves in their environment to overcome interference and enhance wireless communications. This paper considers RIS enhanced D2D communications underlaying unmanned aerial vehicle (UAV) networks with non-orthogonal multiple access (NOMA). The objective is to maximize the sum rate of NOMA D2D communications by simultaneously optimizing the power budget of D2D transmitter, NOMA power allocation coefficients of D2D receivers and passive beamforming of RIS while guaranteeing the quality of services of UAV user. Due to non-convexity, the optimization problem is intractable and challenging to handle. Therefore, it is solved in two parts using alternating optimization. Simulation results unviel the performance of the proposed RIS enhanced D2D communications scheme. Results demonstrate that the proposed scheme achieves 15\% and 27\% higher sum rates compared to the fixed power D2D and orthogonal D2D schemes.
This paper proposes an energy-efficient RIS-assisted downlink NOMA communication for LEO satellite networks. The proposed framework simultaneously optimizes the transmit power of ground terminals of the LEO satellite and the passive beamforming of RIS while ensuring the quality of services. Due to the nature of the considered system and optimization variables, the energy efficiency maximization problem is non-convex. In practice, obtaining the optimal solution for such problems is very challenging. Therefore, we adopt alternating optimization methods to handle the joint optimization in two steps. In step 1, for any given phase shift vector, we calculate satellite transmit power towards each ground terminal using the Lagrangian dual method. Then, in step 2, given the transmit power, we design passive beamforming for RIS by solving the semi-definite programming. We also compare our solution with a benchmark framework having a fixed phase shift design and a conventional NOMA framework without involving RIS. Numerical results show that the proposed optimization framework achieves 21.47\% and 54.9\% higher energy efficiency compared to the benchmark and conventional frameworks.
In this paper, we study a RAN resource-slicing problem for energy-efficient communication in an orthogonal frequency division multiple access (OFDMA) based millimeter-wave (mmWave) downlink (DL) network consisting of enhanced mobile broadband (eMBB) and ultra-reliable low-latency communication (URLLC) services. Specifically, assuming a fixed set of predefined beams, we address an energy efficiency (EE) maximization problem to obtain the optimal beam selection, Resource Block (RB), and transmit power allocation policy to serve URLLC and eMBB users on the same physical radio resources. The problem is formulated as a mixed-integer non-linear fractional programming (MINLFP) problem considering minimum data rate and latency in packet delivery constraints. By leveraging the properties of fractional programming theory, we first transform the formulated non-convex optimization problem in fractional form into a tractable subtractive form. Subsequently, we solve the transformed problem using a two-loop iterative algorithm. The main resource-slicing problem is solved in the inner loop utilizing the difference of convex (DC) programming and successive convex approximation (SCA) techniques. Subsequently, the outer loop is solved using the Dinkelbach method to acquire an improved solution in every iteration until it converges. Our simulation results illustrate the performance gains of the proposed methodology with respect to baseline algorithms with the fixed and mixed resource grid models.