key: cord-0139441-qnij4wio authors: Shih-Chun, Lin; Chia-Hung, Lin; ChuLiang, C.; Shao-Yu, Lien title: Towards Resilient Access Equality for 6G Serverless p-LEO Satellite Networks date: 2022-05-17 journal: nan DOI: nan sha: 2508106dd2132902b6048032ebd629ff92587e96 doc_id: 139441 cord_uid: qnij4wio Low earth orbit (LEO) mega-constellations, integrating government space systems and commercial practices, have emerged as enabling technologies for the sixth generation (6G) networks due to their good merits of global coverage and ubiquitous services for military and civilian use cases. However, convergent LEO-based satellite networking infrastructures still lack leveraging the synergy of space and terrestrial systems. This paper, therefore, extends conventional serverless cloud platforms with serverless edge learning architectures for 6G proliferated LEO (p-LEO) satellite ecosystems and provides a new distributed training design from a networking perspective. The proposed design dynamically orchestrates communications and computation functionalities and resources among heterogeneous physical units to efficiently fulfill multi-agent deep reinforcement learning for service-level agreements. Innovative ecosystem enhancements, including ultrabroadband access, anti-jammed transmissions, resilient networking, and related open challenges, are also investigated for end-to-end connectivity, communications, and learning performance. Emerging proliferated low earth orbit (p-LEO) constellations [1] - [6] with both government and commercial satellites have been regarded as the most promising remedy to provide global coverage and ubiquitous wireless services, bridging the ever-existent digital divide via their global footprints. These broadband connectivities to the Internet allow people to access information and essential services, such as governmental and health systems, ubiquitously without suffering geographical limitations. The need for resilient access equality via LEO satellites' high communications capacity with ultra-wide ranges [5] , [6] is more urgent than ever, as we have abruptly switched to remote living in past years due to the COVID-19 pandemic. Meanwhile, distributed machine learning (ML) brought attentions to decentralized data sources. Such learning technology will likely address multi-dimensional resource allocations for integrated mega-constellations and the sixth generation (6G) networks [2] , [7] , [8] . However, the recent federated or collaborative satellites mainly work on their feasibility via algorithm implementations (e.g., [3] , [4] ). There is little investigation into the redundancy and tradeoff between computations and communications and the dedicated resource orchestrations to realize timely edge learners with efficient data processing. Also, few solutions exist to comprehensively evaluate distributed training performance concerning nonterrestrial networks' peculiarities, e.g., satellite access and multi-tier connected infrastructure. Architectural, management and operational changes are required to realize the ecosystems. This article presents a serverless software-defined networking (SDN) architecture that dynamically orchestrates communications and computation resources for a diverse set of 6G service-level agreements (SLAs). It provides a multi-tier ML framework that uses a unified control platform to optimize networking and resource configurations according to space and ground tiers' interactions. This coherent framework, coupled with underlying software-defined infrastructure, focuses on practical constraints and peculiarities in each tier, such as heterogeneous computing capabilities of ground terminals and size, weight, and power (SWaP)-limited satellites, frequentlyhandover satellite access, and coverage-limited ground tier. Remarkably, the intelligence within multiple ML-SDN control engines (e.g., ground terminals and satellites with computing capabilities) can realize efficient broadband access for ground users concerning software-reconfigurable LEO satellites, datadriven approaches for unknown environments, and different decision timescales of each unit. The design can use multitier ML models to establish high-throughput, reliable end-toend transmissions for global connectivity. It is noteworthy that our designs tightly align with the latest industry specifications. For instance, 3GPP Release 17 considers satellite mobility at different orbital heights to support non-terrestrial networks with 3GPP NR on the ground. Release 18 in 2022 creates 5G Advanced to include new intelligent, ML-enabled solutions to boost mobile broadband and verticals performance. The proposed architecture and solutions will bring several benefits to 6G p-LEO satellite ecosystems, as follows: also be transferred to its successor, preventing always training from scratch for SWaP-limited devices. 3) We enable data-driven multi-user access control for the ground-space eco-network. Serverless computing architectures provide infrastructure controllability to the multi-tier ML framework. This framework can establish efficient ultra-broadband sensing and communications between the two tiers to satisfy 6G requirements. 4) We achieve reliable software-defined internetworking through spectrum harmonization and hyper-connectivity. New networking designs are realized to address heterogeneity, scalability, performance, and reliability fully. For example, we investigate (i) anti-jammed multiple-input-multiple-output (MIMO) p-LEOs with Ka-bands, MIMO beamforming, and frequency-hopping techniques, (ii) "Space Highway" to leverage megaconstellation structure and enhance global transmission capacities, and (iii) cyber-harden opportunistic multipath routing to solve long delay and cyber-attacks. Therefore, our innovations significantly enhance end-to-end performance and impact future human society in isolated or remote communities and landlocked areas with limited infrastructure investments. Our work will also facilitate distributed deep training development with fast-adaptiveness and efficient multi-tier processing while tackling non-terrestrial system heterogeneity and dynamics. The rest of this article is organized as follows. The following section gives the state-of-the-art. We then present serverless edge architectures and discuss ultrabroadband access and resilient networking. The final section concludes the article. To integrate p-LEO or satellite swarm with ground communications, recent studies [1]- [8] focus on spectrum sensing and sharing as well as ML-based resource allocation for such non-terrestrial coexistence. In [3] , cognitive radio technologies are adopted to detect the channel state of primary signals and suppress co-channel interference for CubeSat swarm networks. In [4] , beamspace MIMO is exploited for downline satellite swarms, requiring only position information for distributed linear precoders and a ground equalizer. Authors in [6] summarize ultra-dense LEO satellite networks and introduce satellite access architectures with supporting technologies and use cases. Furthermore, SDN-based management [7] provides a deep Q-learning approach to orchestrate networking, caching, and computing resources jointly for satellite-terrestrial networks. In [1] , optimal network control structures are studied to improve the temporal control effectiveness with the least number of controllers. Authors in [2] further consider threedimensional (3D) terrain surface coverages by designing hierarchical unmanned aerial vehicles (UAVs) swarms via deep reinforcement learning algorithms. Moreover, an automatic network slicing platform for the Internet of space things is presented in [5] , which carries out various SLAs over the space-ground integrated infrastructure. Also, multi-user access schemes to non-terrestrial bases stations are investigated in [8] via deep reinforcement learning to provide high throughput and fewer handovers for 6G traffic. Still, sensing-enabled coexistence, resilience access, and resource management developments for 6G p-LEOs are in infancy and require innovations and new approaches. A. ML-SDN Control Engines and Infrastructure Fig. 1 shows the proposed serverless edge architecture in ground and space tiers for 6G p-LEO satellite networks. The ground tier consists of several terrestrial systems, such as the Internet of things (IoT), UAVs, vehicle-to-everything (V2X), etc.; each system has a dedicated ML-SDN control engine. These control engines integrate SDN controllers with ML algorithms and manage computing, storage, and communications resources. They receive resource and service requests and training data from the serving systems and, in turn, assign tasks and control decisions back. The space tier includes p-LEO satellites from different operators, such as SpaceX, Amazon, space development agency (SDA), in orbits, and operating systems. A serverless edge platform is established for a scalable and unified control plane to coordinate multi-tier engines and constitute a shared resource pool for virtualization and network slicing. It can alleviate the disturbance to physical infrastructure units giving time-varying resource availability and heterogeneity. Specifically, the serverless platform pushes the controllability to network edges via local control engines with edge and serverless computing. As 6G p-LEO systems cover an ultra-wide geographical and spatial area, a single centralized control with huge decision parameters is not a feasible solution. TABLE I summarizes three control plane implementations, corresponding attributes, challenges/costs, and learning deployments for multi-tier ML-SDN engines. These multiple engines can be organized in a fully distributed, multi-domain flat, or multi-layer hierarchical manner to provide control scalability and boost learning efficiency. Notably, the multi-domain and multi-layer control can realize optimal global learning by splitting enormous optimization dimensions into collaborative engines' tasks. Such collaboration can be revamped from SDN's east-bound and west-bound application programming interfaces (APIs). The south-band APIs are expanded to help ML-SDN engines to communicate and control the respective underlying physical resources. The expanded designs are dedicated to systems' various functions, such as satellite beam steering, UAV movement control, and resource block allocation in base stations. The proposed 6G serverless platform aims to establish an intelligent networking architecture that effectively manages computation and communications resources to meet all SLA demands in 6G p-LEO systems. As in Fig. 1 , this platform adopts a Function as a Service (FaaS) and orchestrates function containers among (geographically distributed) ML-SDN control engines for highly flexible virtualization infrastructure. It provides policy-based guidances for ML workflows and consists of two crucial modules. First, the service management module enables diverse p-LEO and ML applications to request SLA portfolios (e.g., link throughput, end-to-end latency, etc.) from the lower management module. Each application is decomposed into functions (e.g., service discovery, deployment, scheduling, caching, etc.) and defines a dedicated service function chain upon the infrastructure. Second, the network and resource management module collects global network status, allocates multi-dimensional resources, and regulates multiple control engines with the ML deployments. Specifically, the network management platform maintains the topology, decides the best routes, and schedules and monitors the upper function chains. The resource orchestration platform allocates and controls underlying resources from distributed infrastructure. With the above designs, the 6G serverless platform can efficiently explore network adaptivity and realize optimal space-ground policies at serverless edges through network automation. One primary objective of this serverless platform is to dynamically adjust ML workloads among heterogeneous computing units while considering each of their real-time capabilities. As control engines can equip with multi-agent learners, different ML deployments in TABLE I can be implemented to ensure effective engine coordination for outstanding system performance. Specifically, a multi-agent deep reinforcement learning (MADRL) algorithm has three design components: observation processors, action predictors, and deep reinforce-ment learners (DRLs). An agent interacts with the environment (i.e., p-LEO networks) and obtains local observations and models (e.g., current state, action, reward, and next state). An action predictor can characterize agent collaboration by predicting other agent's actions. DRLs can derive the optimum policies based on the observation and prediction results. Furthermore, each agent/engine can also upload its local parameters to the serverless platform, training neural networks with global information. Thus, as in Due to recent advances in transceiver hardware, frequencyagile ultra-broadband reconfigurable frontend is envisioned to realize full-spectrum (1 GHz to 10 THz) sensing and communications that meet data rate, reliability, and scalability requirements. For example, the National Aeronautics and Space Administration's (NASA's) Aura satellites collect radiometric data on 118 GHz, 190 GHz, 2.5 THz, etc., in their remote sensing and earth exploration services. In 5G NR release 17, frequency ranges 1 (sub-6 GHz) and 2 (millimeter wave, mmWave) are considered to support cellular V2X (C-V2X) communications and enhance wide-area coverage. Hence, the spectrum innovation technology and multi-antenna access will need to fast, agilely, and automatically utilize the ultra-broadband to optimize physical transmissions. The network infrastructure and satellite terminals can recognize spectrum usages and exploit radio resources for communications efficiency through deep reinforcement learning based on serverless edge architectures. All-spectral analytic for intelligent sensing and autonomous configuration thus becomes a prerequisite of optimizing spectral efficiency in 6G p-LEO networks. The main technical challenge is to develop innovative spectrum sensing techniques and sensing-informed communications for dynamic access to all-spectral resources. First, evolving from MHz to GHz spectrum characterization, wireless learning features (e.g., signal waveforms, cyclic spectrums, complex correntropy) should be extracted from raw sensory input. Accordingly, deep learning-based wideband sensing techniques (e.g., wavelet detection or compressed sensing) can be designed to identify available spectrum efficiently. Such real-time learning variants (e.g., real-time inference, fast spectrum analytic) under practical wireless channels (e.g., fast fading for highly mobile vehicles) can be further investigated for timely sensory processing. Next, these sensing algorithms can be extended with end-to-end learning techniques concerning hardware impairments, physical-layer model mismatches, and nonlinearities. An end-to-end spectrum learning with vehicular channel awareness [9] is proposed to disclose the usage behaviors of THz bands fully. Fig. 2 shows the performance of ultra-broadband spectrum recognition and sensing for a recent generative adversarial network (GAN)-based solution [10] and our work [9] . We set up a practical transportation environment from downtown Raleigh, North Carolina, and C-V2X systems with co-existing sub-6 GHz, mmWave, and THz communications. The results imply that our scheme can effectively extract and learn multiple simultaneous connections and outperforms the GAN realization for all bands by jointly designing spectrum compression and reconstruction. Moreover, deep recurrent learning, e.g., long short-term memory and gate recurrent unit, can be employed to develop dynamic all-spectral access based on time-series results from ultra-broadband sensing. Learning-based spectrum decision, sharing, and mobility can be proposed to avoid radio interference and optimize shared spectrum allocation by considering delayed sensing data. Besides, to enable radio components adaptive to time-vary environments, an autonomous frontend setting should be designed to tune channels and transmission power levels in real-time. Notably, from the designed allspectral sensing and communications, comprehensive frontend configurations (e.g., analog electronics, bandwidth sensitivity, position) can be automatically and optimally adjusted by sensory processing and training-driven decision making (e.g., reinforcement learning). Two crucial challenges are supporting timely beamforming with high bandwidth for multi-antenna ultrabroad systems. First, supervised deep learning can address complicated beamformers with large antenna arrays (e.g., perfect channel state information via channel estimation). However, it is limited by the performance of labelling algorithms, which must label vast input data (especially in multi-user scenarios) during offline training. Second, hybrid beamformers need low-latency beam management to constantly provide good transmission quality in fast time-varying channels. Such channel conditions are due to peculiar LEO satellite movement and communications band (e.g., mmWave's blockage sensitivity, THz's pronounced molecular absorption, and spreading losses). Unsupervised reinforcement learning-based beamforming can cope with these challenges by providing fast beam tracking for the net multi-antenna gain. In serverless edge architectures, satellite swarms can coordinate for joint operations to act as a distributed antenna system. The time-varying MIMO fading channel on space-ground communications can then be formulated by extending the K-user interference channel. The average downlink data rate can be obtained via medium access control (e.g., asynchronous direct-sequence code-division multiple access). Unsupervised learning requires "differentiable" objective functions. So, we adopt a constrained user sum-rate maximization for the beamforming design. This framework will optimize power and spectrum allocations and beamformer matrices subject to maximum available power and spectrum constraints (i.e., from automatic spectrum analytic) and stochastic computation delay requirements. The primary objective is to introduce computation-efficient unsupervised algorithms that consistently provide good beamformers adaptive to timely environmental changes. But, model-based schemes are computationally prohibited in real-time. Recent deep learning beamforming employs regularized loss functions or scaled beamforming output to satisfy power constraints, having a performance gap compared to the weighted minimum mean square error (WMMSE) counterpart. Deep unfolding techniques have been used to leverage optimization designs and residual neural networks and develop two consecutive modules for efficient unsupervised beamformers [11] . First, a coarse estimator module is proposed to address the constrained maximization framework for outperforming WMMSE in sumrate. Then, gradient descent beamforming is empowered with a deep unfolding module to enable fast convergence with superior performance. A balance between system resilience and rapid beam alignment/steering can also be established by upgrading unsupervised beamformers with reinforcement learning-enabled tracking. Multi-agent Q learners will be adopted with received power levels to develop new beam tracking to guarantee computation delay requirements. V. ROBUST AND RESILIENT SATELLITE NETWORKING FOR SERVICE-LEVEL AGREEMENT ASSURANCE This section exploits the new spatial dimension of p-LEO for resilient end-to-end networking. We consider anti-jamming capabilities for space-ground communications and investigate novel space highway and cyber-harden transmissions. An adaptive power control mechanism can be developed in higher bands (e.g., Ka at 26.5-40 GHz) to realize efficient ground-LEO access (particularly for uplink communications as 6G backhaul or integrated access and backhaul). This mechanism minimizes the transmit power by jointly employing satellite channel estimation and a feedback control loop. Ka-band systems (currently applied by Starlink by SpaceX) deliver substantially greater throughput than previous Kuband offerings. Fig. 3 shows downlink packet error rates for an LEO satellite to a line-of-sight ground terminal. Ten terminals share the channel via direct sequence spread spectrum (DSSS) with 240 (i.e., 23 dB) spreading factor, Raised Cosine pulse shape with 0.35 roll-off factor, 16 Mbps data rates, and 1,500 bits packet size. The results imply that a 12 dB gain can be obtained with turbo codes and nominal equivalent isotropically radiated power (EIRP). Also, from SDA optical communications terminals, the modulation types use on-off-keying non-return-to-zero and m-ary pulse position modulations with radio resources located at 193.1 and 195.1 THz. Hence, uplink channel estimation can be employed to enable adaptive power control schemes based on physicallayer specifications and downlink performance. For example, an auto-regressive moving-average model generally describes the dynamics of rain fading in Ka-bands. The transmit power can then vary following the fading gain to keep the signal-tonoise power ratio (SNR) level while maintaining the requested power outage probability. Furthermore, due to the new spatial dimension in MIMO p-LEOs, such a satellite swarm can utilize line-of-sight channels and ultra-wide coverage for more robust communications with higher throughput. These MIMO LEOs adopt MIMO and spread spectrum technologies to prevent active jamming attacks or harmful co-channel interference. For example, antenna beamforming can eliminate interference by directing a null toward a jammer. In [12] , a DSSS system is revamped and tested experimentally for above 100 GHz and secure spectrum sharing, ensuring coexistence between ground THz active Moreover, the anti-jamming capability of MIMO p-LEOs can be increased to the next level by exploiting their spatial dimension with MIMO beamforming and frequency hopping spread spectrum. In particular, uplink/downlink transmissions can smartly allocate specific "virtual" rays among total available rays through p-LEO coordination to avoid jamming signals. Larger-scale p-LEO coordination can be accomplished via hardware synchronization, timing protocol designs, or even software-defined architectures, like serverless edges. Thus, massively networked MIMO techniques can be established in satellite networks to achieve various SLAs (e.g., variable-rate MIMO links, robust and bandwidth-efficient communications, etc.). Due to serverless edge infrastructures, heterogeneous satellite constellations from different operators, such as governments or commercial companies, can now integrate into a hybrid space architecture through space-based adaptive communications node (SBACN) and API development for cross-constellation communications command and control. As shown in Fig. 4 (a) , a hierarchical structure can be established where an SBACN center connects to several SBACN terminals. These terminals integrate existing satellite operators' constellations by acting as their gateways. SBACN terminals only need to update the resource information and service requests to the SBACN center via serverless edges and enable FaaS for connectivity and SLA satisfaction without configuring or managing resources. On the other hand, the SBACN center coordinates reliable inter-constellation-level connectivity, while each operator handles its intra-constellation-level management. The SBACN center hosts on-demand applications through dynamically instantiated containers and effectively computes optimal routing-path planning and resource allocation designs, avoiding underutilization while maintaining high throughput. Also, iteratively upgrading efforts for the APIs and underlying algorithms are negligible due to the software-defined nature. Based on the scalable cross-constellation p-LEO platform, we can develop innovative space highway and delay-optimal multipath TCP (MPTCP) for efficient end-to-end networking. In Fig. 4 (b) , globally scattered p-LEOs will run above terrestrial infrastructure with tremendous ground terminals. These satellites assist terrestrial communications by providing short-cut paths among ground terminals. The ground terminals now can use the upper-tier satellite network for ultra-fast data delivery rather than conventional multihop terrestrial transmissions with very-long routes. We describe research challenges for effective space highway designs as follows. Possible topologies for the p-LEOs should be investigated concerning their topological properties (e.g., network diameter, degree distribution, average hop distance) to determine the best topology with minimal latency. From our previous work with geometric random graph [13] , a preferable candidate could be the small-world satellite topology from the small-world phenomenon in social networks. Then, queueing network analysis and statistical quality-of-service provisioning can be applied to assure SLAs. In addition, satellite-assisted communications suffer from long delays and frequent ground-satellite handovers (both are problematic for TCP connections). MPTCP protocols address these challenges by exploiting several Internet paths between a pair of hosts while presenting a single TCP connection to the upper layer. Thus, the upper layer applications only need to deal with a single logical master TCP connection. Multiple sub-flows are running underneath, each of which is a conventional TCP connection. To extend MPTCP to p-LEOs with serverless edge architectures, we provide a centralized delay-optimal MPTCP that minimizes the end-to-end delay with MPTCP while eliminating frequent control message signaling from distributed algorithms. Centralized controllers will facilitate (i) topology and traffic monitoring/ prediction, (ii) MPTCP awareness, and (iii) multiple disjoint paths discovery. In particular, we consider the p-LEO network topology and traffic modeling when formulating a nonlinear optimization to find end-to-end routes with minimum average delay and no congestion. We then exploit fast algorithms to provide the optimal solution within a few milliseconds, thus achieving timely and reliable p-LEO satellite systems. For connectivity robustness and resilience, defensive and self-healing algorithms should be designed to tackle possible cyber-attacks and provide ceaseless connections in node failures, deep fades, and malicious behaviors. A straightforward solution is to create backup paths, but it is resource prohibitive in assigning wireless backup capacity. Hence, we are revamping our opportunistic multipath routing algorithms [14] , [15] to provide cyber-harden p-LEO communications. In [14] , a cognitive and opportunistic relay solution is designed for reliable communications and connections in a machine swarm. This design enables machines to cognize and adapt to environments for mitigating inter-system interference with existing networks and realizes opportunistic selections of cooperative relay machines based on link qualities. We extend with p-LEOs' peculiarities to distributedly concatenate link transmissions for satellite swarm resilience. In [15] , "virtual" MIMO at the session level is established, and probabilistic networkcoded routing is developed for large-scale cognitive machine swarms. This work enables spatial multiplexing and diversity with session-level traffic and ensures end-to-end delay by employing network coding techniques with underlaid routing algorithms. The developed solution expands multi-user MIMO into multihop multipath transmissions. We thus exploit such dynamic cooperation for the fault tolerance of p-LEO systems, robust to node or link failures. Remark (Practical Experimentation). In February 2022, Lockheed Martin has awarded an SDA's Tranche 1 Transport Layer contract to demonstrate an interoperable, connected, secure mesh network of 42 LEO satellites that link terrestrial warfighting domains to space sensors. As deeply involved in this development, we conduct the proof-of-concept of this paper's solutions for anti-jammed p-LEO satellites and crossconstellation communications. Our designs' on-orbit deployment is expected to be in 2024-2026. We also expand Cisco edge AI with our serverless edge framework in telemedicine and broadband connectivity in rural areas. LEO mega-constellations have promised to serve isolated or remote communities and fulfill the needs of landlocked areas with limited infrastructure investments. However, there is still little work that simultaneously addresses the satellite access and inter-tier networking within the 6G context and provides a coherent serverless architecture for such ecosystems. This article introduces unique serverless edge architectures with multi-tier deep reinforcement learners and emphasizes necessary architectural, management, and operational advances, thus bringing a new frontier for resilient access equality. Mega satellite constellation system optimization: From a network control structure perspective Deep reinforcement learning based three-dimensional area coverage with uav swarm Spectrum sensing of cognitive radio for cubesat swarm network Beamspace mimo for satellite swarms Towards automatic network slicing for the internet of space things Ultra-dense leo: Integration of satellite access networks into 5g and beyond Deep q-learning aided networking, caching, and computing resources allocation in softwaredefined satellite-terrestrial networks Deep reinforcement learning for multi-user access control in non-terrestrial networks Tulvcan: Terahertz ultra-broadband learning vehicular channel-aware networking End-to-end deep learningbased compressive spectrum sensing in cognitive radio networks Unsupervised resnet-inspired beamforming design using deep unfolding technique Ultrabroadband spread spectrum techniques for secure dynamic spectrum sharing above 100 ghz between active and passive users Statistical dissemination control in large machine-to-machine communication networks Cognitive and opportunistic relay for qos guarantees in machine-to-machine communications Statistical qos control of network coded multipath routing in large cognitive machine-to-machine networks