Next Article in Journal
Emotion Recognition from Physiological Channels Using Graph Neural Network
Previous Article in Journal
Sparse Damage Detection with Complex Group Lasso and Adaptive Complex Group Lasso
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Reinforcement Learning-Based Resource Allocation for Satellite Internet of Things with Diverse QoS Guarantee

1
Command & Control Engineering College, Army Engineering University of PLA, Nanjing 210007, China
2
Beijing Information and Communications Technology Research Center, Beijing 100036, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(8), 2979; https://doi.org/10.3390/s22082979
Submission received: 21 February 2022 / Revised: 28 March 2022 / Accepted: 1 April 2022 / Published: 13 April 2022
(This article belongs to the Section Internet of Things)

Abstract

:
Large-scale terminals’ various QoS requirements are key challenges confronting the resource allocation of Satellite Internet of Things (S-IoT). This paper presents a deep reinforcement learning-based online channel allocation and power control algorithm in an S-IoT uplink scenario. The intelligent agent determines the transmission channel and power simultaneously based on contextual information. Furthermore, the weighted normalized reward concerning success rate, power efficiency, and QoS requirement is adopted to balance the performance between increasing resource efficiency and meeting QoS requirements. Finally, a practical deployment mechanism based on transfer learning is proposed to promote onboard training efficiency and to reduce computation consumption of the training process. The simulation demonstrates that the proposed method can balance the success rate and power efficiency with QoS requirement guaranteed. For S-IoT’s normal operation condition, the proposed method can improve the power efficiency by 60.91% and 144.44% compared with GA and DRL_RA, while its power efficiency is only 4.55% lower than that of DRL-EERA. In addition, this method can be transferred and deployed to a space environment by merely 100 onboard training steps.

1. Introduction

Satellite Internet of Things (S-IoT) [1] is a wireless communication scenario where the Internet of Things (IoT) terminals transmit data through satellites. IoT devices are sparsely deployed in rural areas, including forest, mountains and the sea, with low financial efficiency. Due to the geographical limitation, it is unrealistic to deploy a Narrowband Internet of Things (NB-IoT) network in the ocean or mountain areas. Satellite constellations can be a promising plan to serve IoT devices in remote rural areas. For instance, satellite constellations such as Orbcomm and ARGOS have already played an indispensable role in the industry application of remote areas, including mining, oil and gas exploitation, forest fire prevention, disaster prediction, and environment protection.
Despite its promising prospect, S-IoT still faces technical challenges posed by not merely the limitations of satellite communication but by traits of the IoT scenario. This paper focuses on the resource allocation problem [2], whose challenges can be listed as follows.
  • To satisfy the requirements on delay, reliability, the bandwidth of various IoT applications, diverse Quality of Service (QoS) guarantee [3] is a major issue for transmission mechanisms in IoT systems. More specifically, various types of services in S-IoT make it necessary for its resource allocation algorithm to pay attention to diverse QoS requirements of each data transmission.
  • Compared with conventional human user terminals, the IoT terminals located in remote areas are usually battery powered and are thus energy limited, whose energy consumption of data transmission is usually inversely proportional to their lifetime. Therefore, balancing energy saving and transmission rate promotion is necessary when allocating IoT terminals’ transmission power resources.
  • Compared with terrestrial IoT networks, S-IoT achieves wider coverage and supports massive IoT terminals. As the computational complexity of most resource allocation algorithms is substantially proportional to the number of users, data transmission of large-scale terminals [4] imply challenges for dynamic resource allocation.
  • In a S-IoT scenario, it is always impractical to obtain perfect global Channel State Information (CSI). First, the channel quality of S-IoT, whose spectrum band is usually on Ka, is sensitive to weather conditions, especially to rainfall and snow, leading to a dynamic transmission environment of S-IoT. Due to the long satellite transmission link, the reported channel quality information tends to expire due to delay. Moreover, there are still considerable errors in the channel estimation methods despite the effort of researchers. As a result, the aforementioned errors in CSI should be considered in resource allocation.
The challenges mentioned earlier have already captured the attention of scientific communities. Although much work on resource allocation has examined how to ensure QoS requirements, deal with time-varying channel quality, save energy, and serve largescale terminals, most of the existing research has the following limitations.
  • Most of the existing QoS-guaranteed resource allocation methods concentrate on a specific type of QoS metric and is implemented on homogeneous terminals in a local area. Nevertheless, considering the tens of kilometers of satellite coverage, resource allocation methods in S-IoT should cope with diverse QoS requirements of massive amounts of terminals.
  • Quite a few researchers have proposed energy-saving resource allocation methods, whereas they usually focus on the energy efficiency of satellites in a downlink scenario rather than the energy-saving issue of the terminals.
  • The existing model-based methods, limited by the paradigm of optimizing a specific communication model, are difficult to satisfy the intermittent access of massive terminals. The complexity of the optimization algorithm increases exponentially with the scale, resulting in difficulties in the real-time allocation of large-scale terminals. Conversely, data transmission of remote IoT terminals is usually intermittent with a limited duration, showing that the transmission resource application arrives in an online manner. However, the model-based methods need to set a fixed number of user terminals in the initialization process and are thus incapable of such online problems.
  • The model-based methods usually build on specific channel model assumptions or use accurate global CSI as the input of methods. For the first set of methods, when the real-world channel is different from the predefined channel model due to interference or weather changes, their performance may severely degrade error. For the first set of methods that include channel quality, noisy and outdated CSI will lead to an accumulation of errors.
With regard to the limitations of model-based methods, an intelligent method based on deep reinforcement learning (DRL) has been introduced to the resource allocation field [5]. The essential idea of DRL is in making decisions based on the observation of the dynamic environment and adjusting the strategies according to environmental feedback, indicating that DRL-based resource allocation methods can adjust their strategies with the variation of channel quality [6]. Furthermore, the DRL, professionally designed to solve sequential decision problems, is naturally suitable for the online resource allocation problem and tends to promote the sum of long-term system rewards, rather than immediate revenue. If the method only focuses on short-term objectives, it may fall into local optima. Such long-term optimization, which is accomplished by discount factors and the updated formula of the value function, is the main advantage that makes DRL successful in sequential decision problems.
Taking advantage of DRL, we propose an online energy-saving uplink resource allocation method with diverse QoS guarantees for S-IoT. The proposed method in this paper is the first to simultaneously address diverse QoS constraints, massive terminal burst data transmission, and energy efficiency of terminals. The method adopts DRL to construct an online resource allocation pipeline for large-scale terminals. More specifically, for each emerging transmission request, the intelligent agent collects its contextual information and then simultaneously allocates channel and transmission power. The agent can promote long-term energy efficiency for S-IoT by learning from the feedback of the actual environment to guarantee diverse QoS.
The main contributions of the proposed method can be summarized as follows:
  • The resource allocation problem in S-IoT was modeled to promote long-term energy efficiency with various QoS requirements. Taking the energy efficiency of remote IoT terminals into consideration prolongs the S-IoT terminals’ lifetime. Furthermore, QoS requirements include the terminals’ diverse requirements on delay, reliability, and transmission rate rather than on a certain QoS constraint. This model is more suitable with the S-IoT scenario where heterogeneous IoT applications exist in considerably broad coverage of the satellite.
  • The massive terminals’ resource allocation problem is formulated as the Markov Decision Process (MDP). In response to currently generated terminal requests of each time stick, such an online resource allocation framework is more consistent with the actual practice of IoT systems where terminals always transmit data intermittently. In addition, this method’s computational complexity remains unchanged with the increase in IoT terminals, as this pipeline only concerns the currently generated requests, regardless of the other existing IoT terminals.
  • A DRL-based method is proposed to solve the diverse QoS resource allocation problem. The agent observes contextual information, including channel quality, data transmission amount, terminal location, and QoS requirement to make an online channel and power allocation decisions. The agent can learn from the feedback of the environment and adjust its strategy when the channel quality or transmission traffic changes. The learning process of the agent does not depend on channel model assumptions or accurate global CSI.
  • A deployment method based on a transfer learning mechanism was proposed to facilitate implementation in a large-scale LEO satellite constellation. By fixing the first several convolutional layers and fine-tuning the last layers, the converged DRL network in the simulation environment can be efficiently transferred to the actual space environment, thus reducing the computation expense and promoting system efficiency considerably.

2. Related Work

Large-scale resource allocation for massive IoT devices is one of the challenging issues that has attracted much attention in S-IoT [7]. Furthermore, De Sanctis et al. provided an overview discussion on QoS management and resource allocation [8].
Several works have focused on energy-saving issues in the research field of S-IoT resource allocation. Considering the limited onboard energy of LEO in the S-IoT scenario, Zhao et al. addressed the energy-saving channel allocation problem with battery load constraints by adopting DRL and taking the normalized power efficiency and service blocking rate criteria as reward function components [9]. Li et al. poured their main attention into the energy-limited remote IoT terminals in rural areas rather than on the energy-saving issue of satellites [10]. This paper adopted unmanned aerial vehicles (UAVs) as relays and then proposed an energy-efficient model by jointly optimizing channel selection, uplink transmission power control, and UAV relay deployment.
QoS is as critically vital as the energy-saving issue in S-IoT, which has been discussed from different views and approaches in recent decades. Transmission rate and delay are frequently considered QoS requirements for bandwidth and delay-sensitive services, such as environmental management. Jia et al. proposed a delay constraint joint power and bandwidth allocation algorithm by analyzing the interactions among inter-beam interference, delay factor, channel conditions, and traffic demand [11]. Furthermore, Liu et al. focused on a NOMA-based satellite industrial IoT system and proposed a power proportion optimization method for beams and nodes to guarantee the QoS, namely transmission delay and transmission rate [12]. Different from the aforementioned deterministic QoS constraints, effective capacity, which holds statistical QoS guarantee, is a more feasible alternative of QoS requirement with the consideration of the variable quality of the satellite channels. Considering the delay requirements of S-IoT devices, Yan et al. employed effective capacity to express delayed QoS requirements and developed a dynamic power allocation strategy for NOMA by the DRL algorithm [13]. Power allocation factors for NOMA users are selected dynamically by DRL in each time stick to maximize sum effective capacity while meeting each user’s minimum capacity demand constraint. Although this method solved the problems of delay and transmission rate constraints, it ignored network reliability indicators, which are of decisive importance for control applications such as position reporting in vessel navigation. Another limitation of existing studies is that they are designed to handle specific QoS requirements for applications of the same type. However, there are usually IoT devices of heterogeneous applications, and thus with diverse QoS demand, to access the same satellite in the S-IoT scenario.
As an essential factor influencing communication performance, CSI has constantly captured researchers’ attention in the resource allocation field of S-IoT. However, in most existing research, CSI is usually treated as a random variable consisting of large-scale and small-scale fading.
Therefore, it is difficult to adjust the resource allocation strategy according to the specific channel quality of a certain moment. There are three alternative ideas to solve this problem. One is allocating resource flexibly according to a specific channel model, which is adopted by Jia et al. [11] Nevertheless, they had to repeatedly build the model and optimize the problem in case the channel quality model changes [14,15], which may be caused by weather variations or external interference. The second solution, adopted by Sun et al., is leveraging a deep neural network (DNN) to approximate the SIC decoding order in NOMA-based S-IoT since the queue state and channel state continually changes [16]. Although such DNN-based methods are efficient and accurate, their accuracy depends on generating training data with the same distribution as real-world data, which is an arduous task.
Furthermore, DRL is the third method to tackle dynamic channel quality. Hu et al. first introduced DRL to the satellite resource allocation and proposed a dynamic channel allocation method for GEO satellites to decrease long-term blocking probability and improve spectrum efficiency [17,18]. Then, a multi-agent reinforcement learning-based bandwidth allocation of each beam was presented to satisfy the varying traffic demand [19]. Zhang et al. studied power allocation and drew support from DRL to adjust each beam’s transmit power according to the varying traffic demands in cache queue and channel conditions [20]. In summary, the success of the above DRL-based spectrum and power allocation methods can be attributed to the interference management and dynamic environmental perception ability of DRL. DRL-based methods [21] allocate resources according to the current specific channel quality and traffic demand, which is similar to optimization or DL-based methods. Furthermore, DRL-based methods can continuously adjust the allocation strategy intuitively to promote long-term reward according to the environment’s feedback. Such characteristics provide them with the following two advantages over DL or optimization-based ones. The first is that DRL-based methods can perform real-time adjustments with dynamic channel quality, which is changing continuously. Although the channel quality is also regarded as the observation of the environment, the DRL method uses DNN to perceive the environmental information, which shows strong robustness to noisy and error data. Therefore, it does not completely rely on the accuracy of CSI to ensure performance. The second is their long-term reward promotion of online multiple decision problems, which is precisely needed in large-scale S-IoT terminal access. However, the DL approach focuses more on current benefits, which may lead to a local optimum when solving online problems.

3. System Model and Optimization Formulation

3.1. System Model

This paper explores an uplink scenario where the multibeam LEO constellation provides services for heterogeneous remote IoT terminals with various QoS requirements. As shown in Figure 1, each multibeam of LEO satellites provides data transmission services for remote IoT terminals within its beam coverage, such as forest fire monitoring sensors and hydrological sensors. As network access selection has not been studied in this paper, it is assumed that all IoT terminals access their nearest satellite. Therefore, for IoT terminals, it can be assumed that there is only one satellite above it, regardless of other satellites in the LEO constellation. LEO satellites share the spectrum of the Ka-band and implement a direct forwarding mechanism. Furthermore, another assumption in this scenario is that the terminal remains stationary and is equipped with a single antenna.
Multiple power amplifiers on satellites receive uplink signals from IoT terminals, which are located in the geographical area of multiple corresponding beams. Let K and N denote the number of IoT terminals and beams, respectively. Consequently, the sets of IoT terminals and beams can be denoted by U = u k | k = 1 , 2 , , K and B = b n | n = 1 , 2 , , N , respectively. More specifically, the terminal’s beam allocation depends on their location and can be represented by x k = x k , 1 , x k , 2 , x k , N , x k , n 0 , 1 , where x k , n = 1 denotes that terminal u k is located in the coverage of the n th beam and accordingly transmits data to the n th antenna element of the satellite. The maximum forwarding power of a single beam antenna is P B , while the maximum forwarding power of the whole satellite is P t o t a l .
The overall transmission process of u k ’s intermittent uplink data transmission request can be divided into the following processes. Firstly, u k employs a control channel to send access requests in the access process, containing its specific QoS requirements Q o S k and the amount of data D k to be transmitted. Then, the Centralized transmission control unit (CTCU) on the satellite will take charge of channel assignment and power control to allocate specific channel c k , S and power p k , S for u k . Subsequently, u k transmits its data with power p k , S on channel c k , S in the data transmission process. This paper focuses on the joint channel allocation and power control decision in the second step of the process.
For spectrum resources, the total bandwidth is divided into an independent control channel and M data transmission channels according to the frequency division multiple access (FDMA) paradigm. The available channel set for transmission can be denoted as C = c m | m = 1 , 2 , , M , where each channel enjoys the bandwidth of B . Therefore, the channel allocation aims to select a channel from the channel set C for IoT terminal u k , which can be denoted as a channel allocation vector w k = w k , 1 , w k , 2 , w k , M T w k , m 0 , 1 , where w k , m = 1 indicates that the m th channel c m is allocated to IoT terminal u k . Hence the channel allocation result of all the terminals can be expressed as W = w 1 , w 2 , , w M , W K × M .
In terms of power resource, P k denotes the maximum transmission power of the IoT terminal u k . Though power is theoretically a continuous variable, it is usually discretized in practice in satellite communication scenarios to reduce the computational complexity of a satellite control system. We thus assume that transmission power of IoT terminals is selected from the power set of N p power levels, which can be denoted as P s e t = 1 N p P k , 2 N p P k , , N p 1 N p P k , P k . According to the above definition, the uplink power control problem is also resolved in this paper. An appropriate power level has been selected for u k from the power set P s e t . Such result of power allocation can be denoted in vector manner as p k = p k , 1 , p k , 2 , p k , N P T , p k , n P 0 , 1 , where p k , n p = 1 denotes that the transmission power of IoT terminal u k is p k , S = n p N p P k . The power allocation result of K terminals can then be denoted as P = p 1 , p 2 , , p K , P K × N p .
In summary, for IoT terminals, the channel and power allocation matrix, W and P , represent their power and channel allocation results, respectively.
From IoT terminal u k to b th satellite antenna, the entire link gain can be modeled as
G k , b = G k g k , b n ( t ) G b φ k , b n
In Equation (1), G k is the transmit antenna gain of terminal u k , while G b φ k , S represents the receiving gain of satellite antenna, where φ k , S is the angel between terminal u k and the antenna center of b k . Furthermore, g k , b t is the channel power gain and can be given by g k , b t = P L k h k , b t 2 , where P L k is the large-scale fading component and h k , b t captures all time-varying small-scale fading effects. The small-scale fading is composed of multipath fading and atmospheric attenuation of the satellite link, such as gaseous absorption, cloud attenuation, and rain attenuation Those time-varying components are affected by ever-changing weather and atmospheric conditions.
According to the denotations mentioned above, the received Signal to Interference plus Noise Ratio (SINR) [17] of terminal u k can be expressed as
S I N R k = p k G k , b k j = 1 , j k K w j , c k p j G j , b k + σ 2
Consequently, following [19], the transmission rate achieved by terminal u k is given as
C k = B log 2 1 + S I N R k = B log 2 1 + p k G k , b k j = 1 , j k K w j , c k p j G j , b k + σ 2

3.2. Diverse QoS Constraints of Multiple Requests

In remote S-IoT scenarios, requests of heterogeneous devices have different QoS requirements. For stream transmission services, such as continuous data collection and surveillance services in the forest, QoS requirement is mainly on the available data transmission rates, which should exceed data emerging rates to avoid data discarding. Furthermore, monitoring services, such as disaster identification, agricultural machinery control, and industrial monitoring, may have stringent requirements on delay and reliability. As various QoS requirements on transmission rate, latency, and reliability should be simultaneously guaranteed for multiple types of services in remote S-IoT, we analyze the diverse QoS constraints in this section.

3.2.1. Transmission Capacity Requirement

For real-time services such as video surveillance or continuous data collection, their QoS requirement on data transmission rate can be illustrated as
C k C r e q , k
where C r e q , k is the minimum transmission rate requirement of the service s k of terminal u k .

3.2.2. Reliability Requirement

For reliability demanding services, we employ the outage probability as the transmission reliability metric to be guaranteed. With outage threshold γ 0 and tolerable outage probability p k O of service s k , the reliability requirement can be expressed as,
P S I N R k γ 0 p k O
According to [22], reliability constraint in Equation (5) can be transformed into
S I N R k γ eff = γ o ln 1 1 p k O

3.2.3. Latency Requirement

Due to long communication links, satellite networks should not be expected to provide ultra-low delay guarantee within 1 ms, as 5G terrestrial facilities do. However, S-IoT should still provide multiple levels of delay guarantee for various delay-sensitive services. The latency requirement of services is denoted as T r e q , k for terminal u k ’s service s k .
Strictly speaking, different from the transmission rate and reliability requirements, the end-to-end delay of S-IoT data transmission is not only decided by a resource allocation strategy, but by the routing algorithm as well. The intuitive idea to satisfy end-to-end delay requirements is to guarantee each independent process. This paper, therefore, focuses on the latency guarantee of the uplink transmission process between the remote IoT terminal u k and the satellite, which mainly consists of transmission latency and signal propagation latency. The uplink latency, denoted as T k , can be presented as:
T k = T P + L k C k
where L k and C k are the traffic data size and transmission rate of terminal u k , respectively. In addition, T p is the propagation delay of signal transmitting. Decided by the distance between satellite and terminal u k , T p can be approximately regarded as a constant value due to the high satellite orbit.
Thus, the latency requirement can be written as T k T r e q , k , where T r e q , k is the maximum tolerable uplink latency of the service s k provided by terminal u k .

4. Problem Formulation

The S-IoT uplink resource allocation algorithm needs to employ limited channel and power resources to guarantee IoT services’ diverse QoS requirements and promote system performance on the following objectives. Terminal power efficiency needs to be optimized to prolong the terminals’ lifetime. Conversely, the algorithm intuitively needs to maximize the probability of successful data transmission from IoT terminals.
The energy efficiency of terminal u k is defined as:
EE k t = C k t p k , S t
where C k t represents the transmission capacity of terminal u k in the t th time stick, which may fluctuate according to channel quality and co-channel interference of other recently allocated terminals p k denotes the transmit power of terminal u k .
For transmission success index, A k t indicates whether the data transmission request of u k succeeds, which depends on whether the SINR exceeds the threshold.
A k t = 0 , i f S I N R k t < δ t h 1 , i f S I N R k t δ t h
The optimization problem of the QoS-guaranteed uplink power and spectrum allocation for S-IoT terminals is formulated in Equation (10).
o p t . P 1 = max t = 0 T 1 k = 1 K A k t P 2 = max t = 0 T 1 k = 1 K E E k t s . t . k = 1 K p k , S t P t o t a l max , t k = 1 K x k , n p k , S t P B max , t n C j t δ t h , t u j U O n t n p = 1 N p p k , n p = 1 , p k , n p 0 , 1 , t , k U m = 1 M w k , m = 1 , w k , n 0 , 1 , t , k U ( 4 ) ( 6 ) ( 7 )
The optimization objectives P 1 and P 2 in the above problem represent the maximization of the long-term transmission success rate and power efficiency, respectively. The first constraint indicates that the sum of the total uplink power of all terminals should not exceed the satellite’s capacity, as the satellite’s maximum forwarding power limits the sum of the uplink power received by the antenna in the direct forwarding mechanism. Similarly, the second constraint limits the uplink power of all terminals in each beam. The third constraint demands that the current resource allocation result should not interfere with the normal transmission of the existing terminals (i.e., the co-channel interference should not exceed the threshold). The following two constraints denote that only one channel and one power level are allocated to each terminal. The last three constraint items represent the heterogeneous QoS requirement on transmission rate, delay, and reliability, respectively.
The formulated problem is hard to be directly optimized for the following reasons. First and foremost, the transformation from the long-term optimization objective to a series of sub-problems on time slots needs strenuous effort. Second, the channel and power allocation indicators are binary variables, resulting in a mixed-integer nonlinear programming problem with multiple constraints. In addition, the perfectly known real-time CSI, which is the normal premise of the conventional optimization method, is unrealistic in the S-IoT scenario due to the long transmission delay. Consequently, an intelligent model-free DRL-based approach is proposed in this paper to solve the formulated problem, as further elaborated in Section 5.

5. DRL-Based Online Resource Allocation

The original problem in Equation (10) is first formulated through MDP in this section. Afterward, a DRL-based resource online allocation algorithm called DRL-QoS-RA is elaborated, including action space, state space, immediate reward, and operating process. We also propose a transfer learning-based deployment mechanism in Section 5.4 to reduce computational complexity and promote efficiency.

5.1. Algorithm Framework

To address the challenges of massive terminals’ intermittent transmission requests, this paper employs an online pipeline. The resource allocation problem is defined in Equation (10) as a sequence decision-making problem driven by data transmission requests rather than by the conventional maintenance pipeline of resource allocation methods. At each time stick, currently generated terminal requests will be sequentially allocated resources according to the dynamic environment.
Such an online resource allocation problem can be intuitively regarded as a sequential decision problem and further formulated to MDP, in which an agent in CTCU makes a decision to maximize the long-term reward according to the changing environment.

5.2. Mechanism of DRL

Section 5.1 illustrates the necessity and advantage of the online mechanism and further briefly introduces the MDP problem. This section will further elaborate on the essential elements of the MDP problem and DRL method, namely action space, state space, and immediate reward.

5.2.1. Action Space

The intelligent resource allocation agent decides the terminal’s transmission power and channel for S-IoT terminals. More specifically, the joint resource allocation action can be intuitively defined as a = a p , a c , where a p and a c denotes the power and channel allocation action, respectively. Consistent with notation in Section 3, the action space can be denoted as C s e t × P s e t , where C s e t = c m | m = 1 , 2 , , M and P s e t = 1 N p P k , 2 N p P k , , N p 1 N p P k , P k represent the available channel set and transmit power set of the terminals, respectively. As a result, the size of the action space is A = M × N p .
Since the online allocation pipeline (as illustrated in Section 5.1) is adopted to sequentially allocate resources to the newly generated data transmission, each decision only focuses on the terminal of the current request, rather than allocating resources to each terminal. Thus, the action space is defined for the current terminal.

5.2.2. State Space

In addition to the action space, another essential issue in DRL is constructing appropriate state space, which can be divided into the following two steps, choosing the related information and constructing these elements to feature the input of the DNN. Based on the analysis in Section 3, in the S-IoT scenario, the state s t should include the information of channel quality, current arrival terminal u k , t , and the terminal set U t O n , which contains terminals that have been allocated resource and are still transmitting data to the satellite.
The information of the current arrival request u k , t consists of its location c o r d k , QoS requirement Q k t , and its maximum transmission power P k , which can be denoted as info k t = c o r d k , P k , Q k t .
For terminals that are still transmitting data to the satellite, namely k t O n U t O n , the involved information includes their location c o r d k t O n , QoS requirement Q k t O n , and the occupied channel c k t O n and transmission power p k t O n . Therefore, for services k t O n U t O n , the state information is denoted as info k t O n = c o r d k O n , c k t O n , p k t O n , Q k t O n .
Based on the above analysis, the state at time stick t can be expressed as s t = C S I t , info k t , info U t O n . Note that although the agent’s observation of the environment includes CSI, there can be noise and error without the assumption that global CSI is perfectly known.
Then, to make the feature better observed by the DNN, we reformulate the elements of state information from structural representation to tensor manner, as shown in Algorithm 1.
Algorithm 1: Formulate State Information.
Divide all the in-service terminals into M groups as the interference merely occurred between co-channel terminals, as U c m t .
For   m = 0 ,   1 ,   M :
Rank   the   terminals   using   channel   c m   according   to   the   distance   between   them   and   the   current   arrival   request s   terminal   u k t .   Get   the   top   K N   nearnest   terminals   as   U c m t .
If   U c m t :
For   each   terminal   u j U c m t :
Construct u j s   contextual   information   D u j u k t , p j , S t 1 , D k , C s i j , S t 1 , C j t 1 , E E j t 1 , R c j t , R d j t , R r j t   as   feature   vector   s j t ,   where   D i s u j u k t   denotes   the   distance   between   u k   and   u j ,   p j , S denotes   the   transmission   power   of   u j ,   D k   denotes   the   data   transmission   amount   of   u j ,   C s i j , S t 1   denotes   the   data   transmission   of   last   time   stick ,   E E j t 1   denotes   the   energy   efficiency   of   last   time   stick ,   R c j t ,   R d j t ,   and   R r j t   denotes   the   QoS   requirement   on   transmission   rate ,   delay   and   reliability   of   terminal   u j respectively.
Arrange   the   state   information   vector   s j t   of   the   K N   terminals   as   a   matrix   s c m t K N × 9 ,   which   contains   the   contextual   information   on   channel   c m at time stick t.
Construct   the   state   information   matrix   s c m t   of   each   channel   to   tensor   manner   and   get s t s t K N × 9 × M as the input of DQN
Through the above process, the state information s t = C S I t , info k t , info U t O n is formulated as a K N × 9 × M tensor.

5.2.3. Immediate Reward

To maximize the terminals’ long-term energy efficiency with diverse QoS requirements guaranteed, we attach the QoS constraints to the objective of Equation (10) and reconstruct the objective as
O = t = 0 T 1 a 1 k = 1 K E E k t + k U 1 G a 2 C k t C r e q , k + k U 2 G a 3 S I N R k t γ eff + k U 2 G a 4 T k t T r e q , k
G ( x ) is the piece-wise function, whose expression is,
G ( x ) = A , x 0 x , x < 0
The objective is composed of four parts. The first one corresponds to the terminals’ energy efficiency, while the second to the fourth one indicate the penalty of unsatisfied QoS requirement on transmission rate, outage probability, and latency. The purpose of weight a 1 is to balance the promotion of revenue and the penalty of QoS unsatisfaction, whereas a 2 a 4 aim to normalize the penalty parts. This long-term objective can be divided into that of each time stick in the process of online resource allocation, which can be expressed as
O t = a 1 k = 1 K E E k t + k U 1 G a 2 C k t C r e q , k +   k U 2 G a 3 S I N R k t γ eff + k U 2 G a 4 T k t T r e q , k  
To present short-term benefits achieved by making a resource allocation decision for a specific arrival terminal service, we intuitively adopt Δ O t = O t + 1 O t as the immediate reward to inform the agent how much the total goal increases or decreases because of action a t .
Notice that if the action can’t satisfy the QoS requirements of arrival service and existing services, the action will not be carried out, meaning that the resource will not be allocated. The reward of the actions is thus equivalent to
r t = a 1 k = 1 K E E k t + 1       k = 1 K E E k t                                                                                             i f   a t Φ s a t i s f y t k U 1 G a 2 C k t C r e q , k + a 3 S I N R k t γ eff + a 4 T k t T r e q , k     i f   a t Φ s a t i s f y t
where Φ s a t i s f y t denotes the action set which can satisfy the QoS requirement of existing and arrival services at time stick t .

5.3. Process of DRL-Based Online Resource Allocation

Following the conventional training pipeline of DRL, this section presents the training process of the proposed DRL-QoS-RA algorithm. As illustrated in Figure 2, the agent observes environmental information s t , including channel information, QoS requirement, data amount, and existing terminals’ resource occupation. Then based on a certain policy π , the resource allocation action a t is determined, and then the current reward r t is collected, with the environment changing to s t + 1 . Each time the agent goes through the above process, the experience s t , a t , r t , s t + 1 is collected and put into the experience pool, from which the agent can periodically sample experience data and train the policy neural network to optimize long-term benefits. The converged strategy neural network promotes energy efficiency and transmission success rate with QoS requirements guaranteed.
The goal of this paper is to optimize long-term reward.
max π t = 0 T r t
However, since the system operates continuously rather than in a limited time, RL’s classical theory adds the discount factor γ to Equation (14) to make the long-term reward more meaningful. γ 0 , 1 is used to adjust the short-term and long-term impact [23], in other words, how far the agents consider when making decisions. In terms of intuitive perspective, it makes the agent pay more attention to the impact of its action on the near future state. From the perspective of the agent’s training process, T l e n = 1 / ( 1 γ ) can be used to estimate the number of steps that the agent considers in the future when making decisions. For t > 1 / ( 1 γ ) , the discount parameter γ t is almost 0. It means that the agent’s action selection does not consider its influence on the state beyond the T range, which benefits the convergence of RL methods. Therefore, with discount factor γ , the optimization goal is
V π s = E t 0 + T l e n t 0 γ t t 0 r t
, which indicates that the agent optimizes the long-time revenue by optimizing the discounted reward of the future T steps at each time stick. As a result, the larger that γ is, the more steps the agent takes into consideration with more difficulty of the training process. The smaller that γ is, the more the agent pays attention to the immediate interests. The long-term revenue to be optimized can be denoted as:
max π E T t = 0 γ t r t
where E · denotes expectation and π denotes a specific strategy.
For a certain state, the optimization objective is defined as the state value function in RL theory V π s , which denotes the long-term reward of a specific strategy from state s . Owing to Markov property, V π s can be represented as
V π s t = r t + γ s t + 1 P s t + 1 s t , π s t V s t + 1 , π
In RL theory, the state-action value of action a t and state s is also defined as Q s t , a t , which is the accumulated reward if action a t is chosen by the agent.
Q π s t , a t = r t + γ s S P s t + 1 s t , a V π s t + 1
The goal of RL agent is to learn to perform actions to maximize the sum of benefits received in the long term. To achieve this goal, RL methods use different ways to estimate the state value function V π s or state-action value function Q π s , a . For example, Q learning, as a popular method of RL, adopts a Q value table to estimate Q π s , a .
When the state space becomes huge, the value function approximation methods based on table form encounter a dimensional disaster problem and are thus no longer applicable. To solve this problem, a deep neural network is implemented as the mapping function from state to value V π s or Q π s , a . RL methods that approximate value function through a deep neural network are called ‘deep RL’ (DRL), which are still based on the classical theory of RL, but facilitate deep learning to replace the original tabular value function estimation module.
Deep Q-network (DQN) [24] is a popular method of DRL, which improves the Q-learning method by introducing a neural network to estimate the state action value Q π s , a instead of a Q -value table. The deep neural network in DQN is called the Q network, whose parameter of neurons is denoted by θ. The input of the Q network is the state st, while the output of the network is the Q value of each action under the state st. More specifically, the Q value of each state–action pair s t , a t denotes the long-term reward of choosing the action a t in state s t .
The essential idea of DQN is the same with Q learning, which is to determine the Q value of each action under a specific state when the state transition probability P is unknown, so as to obtain the optimal decision. The optimal strategy is to choose the action with the largest Q value.
π ( s t ) = arg max a A Q s t , a
With this strategy, the state value of s t can be denoted as
V π s t = max a A Q * s t , a
The optimal Q value can be obtained by iteratively calculating Equations (17) and (19). More specifically, the estimated value of Q network can be calculated by
y t = r t + γ max a t + 1 A Q ^ s t + 1 , a t + 1 θ
Therefore, the difference between the estimated and prediction value of Q network can be denoted as
L ( θ ) = E y t Q s t , a t | θ 2
This section adopts a gradient decent method to train the Q network, with L(θ) employed as a loss function. The procedures of the DRL-based online resource allocation algorithm (DRL-QoS-RA) are elaborated in Algorithm 2.
Therefore, by learning the Q network, the network can select the action that can maximize the V value according to the current state, which is max π E t = 0 T γ t r t . The optimization of the long-term reward can be realized by making decisions in this way at every moment. The agent pays attention not only to the immediate rewards concerning the current allocation to the current terminal, but also to the long-term impact of its occupied resources.
Algorithm 2: Implementation of DRL-QoS-RA.
Parameter initialization.
Initialize   network   Q   and   Q with random weight.
Initialize   DRL   training   parameters ,   such   as   target   network   update   step   N Q ,   greedy   exploration   probability   ε ,   Buffer   size   N B ,   replay   start   size   N I S ,   and   batch   size   N B S .
Initialize the S-IoT environment parameter.
For   time _ stick   t = 1 , T do
Collect   new   data   transmission   request   into   set   U n e w t
Collect   the   suspended   data   transmission   into   set   U O n t and release the resource occupied by them.
If   U n e w t :
For   each   data   transmission   request   u k t ,
Reformulate   state   information   s t = C S I t , info k t , info U t O n   into   tensor   manner F e a t u r e k , t .
Calculate   a   candidate   action   set   Φ s a t i s f y t
Adopt   the   tensor   feature   F e a t u r e k , t   as   the   input   of   the   action value   network   Q   and   obtain   the   output   of   network   Q   as   v a t = Q s t s t | θ
With   probability   1 ε   to   decide   a t = arg max v a t ,   otherwise   randomly   choose   a t .
If   a Φ s a t i s f y t   :   allocate   resource   for   u k t   according   to   a t .
Collect   environmental   information   s t + 1 ,   then   formulate   to   F e a t u r e k , t + 1   and   calculate   r t
Store   this   experience   F e a t u r e k , t , a t , r t , F e a t u r e k , t + 1   into   experience   pool   D
If   the   sample   of   D   is   more   than   this ,   replay   start   size   N I S :
Sample   N B S   experience   F e a t u r e k , t , a t , r t , F e a t u r e k , t + 1   from   D
Calculate   y t   and   L θ by Equation (20) and Equation (21) respectively.
Update   the   weights   of   Q   by   minimizing   L θ
Every   N Q   steps ,   update   Q ^   using   weight   of   Q

5.4. Deployment Mechanism Based on Transfer Learning

Despite its advantages in learning from the feedback of the environment, the proposed DRL-QoS-RA method still has limitations when deployed in the S-IoT system. The reason can be attributed to the time and calculation expense of the DRL method’s training process, which is further illustrated as follows:
  • If the training process is implemented on the satellite, and the system suffers from the agents’ decision-making errors during the training process, which will result in the failure of IoT data transmission and a waste of system resources. If the training process is implemented in the simulation environment of the ground control center, the accuracy of the algorithm will suffer from the difference between the simulation environment and the actual onboard environment.
  • Each satellite of a large-scale LEO constellation faces a unique environment to learn from. Thus, the computational expense of training can be tremendous if added up in a massive satellite constellation.
Concerning the shortcomings mentioned above, this section embraces transfer learning and proposes a deployment mechanism for DRL-QoS-RA, whose essential idea is that the filters of deep CONV network are learned to observe the environment feature and can be adapted (or transferred) to another similar environment.
As present in Algorithm 3, by fixing the first several CONV layers, the knowledge learned by the agent in the simulation environment can be preserved. Then, by fine-tuning the last fully connected layers, the Q network can adapt to the target environment. Compared with direct deployment, merely fine-tuning the fully connected layers can reduce the computation expense and promote system efficiency.
Algorithm 3: Deployment process of DRL-QoS-RA.
Build simulation environment in ground control center based on the historical channel state, terminal location and IoT request distribution.
Train   the   model   Q F e a t u r e k , t θ to convergence in a simulation environment
For satellite = 0, 1, 2……:
Copy   the   weight   θ   of   the   DQN   Q
Fix   the   conv   layers   in   the   deep   convolutional   network   and   only   fine tune   the   last   few   convolutional   layers   until   the   agent   converges   to   obtain   Q F e a t u r e k , t θ

6. Simulation Result and Analysis

In this section, the S-IoT environment and DRL parameters are introduced first. Then, the results and analysis of transmission success rate and energy efficiency are given. Finally, the performance of the transfer learning-based deployment mechanism is illustrated.

6.1. Experiments Establishment

6.1.1. Scenario Parameters

In this paper, a multibeam LEO satellite is used to simulate the experimental environment of S-IoT, whose detailed parameters are listed in Table 1. The limited backhaul transmission power of the satellite limits the IoT terminals that the system can support.
Furthermore, in the simulation environment, 10,000 IoT terminals are unevenly distributed in the total beams of a single satellite, whose request arrival follows the Poisson distribution with λ times per hour.

6.1.2. DRL Training Parameters

The adopted DQN structure is a convolutional neural network with an input layer, two convolutional layers and two fully connected layers. The conv layers consist of 16 3 × 3 filters and 32 2 × 2 filters. The number of neurons in the hidden layer is 128 and 48, while the ReLu is utilized as the activation function. All other parameters related to the DQN are listed in Table 2. Note that the listed parameters are selected from multiple simulation tests to balance the complexity and performance of the DRL algorithm.

6.1.3. Comparative Methods

To evaluate the performance of the proposed DRL-QoS-RA algorithm, this section compares it with the following methods,
(1)
Genetic algorithm: In the GA-based [25,26] online resource allocation method, the normalized weighted objective. α 1 P 1 + α 2 P 2 is adopted as the optimization objective with the number of parents Np = 200, probability of variation pM = 0.005, crossover probability pC = 0.05, and number of iterations NI = 800.
(2)
DRL-EERA: DRL-based Energy-efficient resource allocation method. DRL-EERA improves the method proposed in [3] by taking power control into consideration, as [3] only considers the channel allocation problem. More specifically, DRL_EERA adopts the state representation and instant reward in [3], while adding power control in action space to allocate transmission power and channel simultaneously. The training parameters of DRL are consistent with the DRL-QoS-RA method, where the deep neural network includes four layers, namely two convolution layers and two fully connected layers.
(3)
DRL-RA: DRL-based resource allocation method. Similar to DRL-EERA, the action design was modified to simultaneously allocate power and channel based on [9]. Moreover, the training parameters and network structure are consistent with DRL-QoS-RA.
(4)
Random method: Power and channel of the current terminal are randomly allocated.

6.2. Convergence Analysis of DRL-QoS-RA Method

Figure 3 shows the changing of reward through a training process when the terminals’ data transmission frequency is five per hour, representing the convergence effect of the methods. The horizontal axis represents the number of transmission requests, while the vertical axis indicates the value of the reward. The shaded area is drawn according to the standard deviation of rt to show its fluctuation.
The replay start size of DRL-QoS-RA is set to 2000. The agent uses random decision mechanisms for the environment exploration and experience gathering when the number of requests is below 2000. As a result, the reward is not significantly improved in this stage. Then, the value of the reward shows significant growth, indicating the improved performance of the algorithm. After the number of requests reaches 8000, the fluctuation of rt is reduced, implying the convergence of DRL-QoS-RA.
Figure 4 shows each methods’ performance on the reward rt. Similar to Figure 3, the horizontal axis represents the number of transmission requests, while the vertical axis denotes reward rt, which is drawn by averaging steps of 100.
First of all, the performance of the three methods based on DRL, namely DRL-QoS-RA, DRL-RA, and DRL-EERA, is similar to that of the random mechanism before the training process of DRL. The reward of the three methods increases with the number of iterative steps.
On the training process, when the request number is between 2000 and 6000, DRL-RA and DRL-EERA take about 4500 training times to achieve convergence, while DRL-QoS-RA only needs 2000 training times. Such an improvement on convergence efficiency can be mainly attributed to the following two points: (1) Compared with DRL-EERA and DRL_RA, DRL-QoS-RA formulates the condensed feature tensor rather than the intuitive location-based zero-padding mechanism, which is used by DRL-EERA and DRL_RA. As a result, the deep network can perceive the critical information more easily and reduce the number of parameters that the deep network needs to train. (2) The state space of DRL-QoS-RA contains more contextual information, such as the amount of data to be transmitted, channel quality, and QoS requirement.
When the request number is more than 6000, all four methods obtain convergence. Compared with the GA method, the reward after convergence of DRL-RA is lower by 14.3%, and the DRL-EERA method achieves similar performance, while the reward of the proposed DRL-QoS-RA method is 11% higher than that of GA.

6.3. Deployment Process Simulation Based Transfer Learning

To demonstrate the adaptability of our deployment mechanism based on transfer learning, we evaluate our method in the transfer learning setting to test the transfer mechanism proposed in Section 5.4. Specifically, the network environment is changed after the 10,000th terminal request, including the quantity and distribution of terminals, their data transmission parameters, and channel quality.
As shown in Figure 5, it takes 4000 times to train DRL-QoS-RA from the initial state to converge. As for the transfer stage, only 1000 times or even about 100 times of training are needed for adapting to the new environment with different distribution and to reach acceptable performance. Furthermore, Figure 5 shows the performance of a transferred neural network. Compared with DRL-QoS-RA trained from the initial state, the transferred one achieves an approximate effect on transmission success rate and power efficiency. To sum up, the transfer mechanism can effectively reduce the onboard training time and computing expense in system deployment.

6.4. Transmission Success Rate and Power Efficiency

The simulation result of transmission success rate is illustrated in Figure 6a, where the horizontal axis denotes the terminals’ average arrival and thus shows the IoT transmission traffic. When the traffic load exceeds a certain limit, the transmission success rate of all methods begins to decrease with the increase in traffic. The DRL-RA, which takes the success of the current transmission as the immediate benefit, achieves the best effect. The proposed DRL-QoS-RA algorithm can also achieve a higher transmission success rate than the DRL-EERA and GA. For example, when the transmission success rate is 0.8, DRL-RA achieves the traffic arrival rate λ = 10.90, whereas DRL-QoS-RA, DRL-EERA and GA can carry the traffic with the arrival rate λ = 8.48, λ = 4.87, and λ = 5.21, respectively. In other words, the proposed DRL-QoS-RA method can improve transmission success rate by 74.12% and 62.76% compared with DRL-EERA and GA, respectively. The reason why DRL-RA achieves better performance than the other methods may lie in the fact that its reward concentrates on success rate.
Figure 6b shows the trend of the terminals’ energy efficiency with the gain of beam traffic. With the increase in the average transmission rate of the terminal, the power utilization rates of DRL-QoS-RA, DRL-EERA, DRL-RA, and GA show a similar trend. After a slight increase, they decrease and become stable eventually. The reason for the subsequent decline stage is that with the increase in requests, co-channel interference inevitably increases, and thus the increase in throughput is less than the power consumption. Then, with the continuous increase in traffic flow, the system becomes saturated by rejecting many new transmission requests. As a result, the power utilization rate of the system appears to be stable. Concerning the power efficiency of S-IoT’s normal operation condition, when we can find that the proposed DRL-QoS-RA is 0.66 Mbit/(Joule), while that of DRL-EERA, GA and DRL-RA is 0.69 Mbit/(Joule), 0.41 Mbit/(Joule), and 0.27 Mbit/(Joule), respectively. The proposed method can improve the power efficiency by 60.91% and 144.44% compared with GA and DRL_RA, while its power efficiency is only 4.55% lower than that of DRL-EERA.
Although DRL-EERA and DRL-RA achieve the best performance in energy efficiency and success rate, respectively, the proposed DRL-QoS-RA better achieves a trade-off between energy saving and transmission QoS satisfaction.
Table 3 shows the success rate, energy efficiency and computational time of the above six optimization algorithms under different transmission frequencies. The calculation time of the GA method increases greatly with the increase in terminal request’s frequency; thus, it is not suitable for S-IoT. The calculation time of the four DRL-based methods remains stable with the increase in in-service terminal. As DRL-QoS-RA transferred adopts the transfer learning mechanism, its training steps are lower than the other three DRL-based methods and can achieve a favorable transmission success rate and energy efficiency with lower calculation cost. Similar to the performance in Figure 6, the proposed DRL-QoS-RA and DRL-QoS-RA-transferred can effectively achieve the tradeoff between energy utilization and transmission success rate. Such a promotion is presented with either low terminal transmission frequency (when S-IoT is relatively idle) or high terminal transmission frequency (when S-IoT tends to be saturated).

7. Conclusions

Aiming to solve the uplink channel allocation and power control problem of large-scale terminals with various QoS requirements in the S-Iot system, this paper proposes the DRL-QoS-RA method for online joint resource allocation based on DRL. Compared with conventional DRL methods, the success of DRL-QoS-RA can be attributed to (1) the comprehensive reward concerning QoS requirement, transmission success rate, and energy efficiency; and (2) contextual information, including location, resource occupation, CSI, and QoS requirement. Furthermore, a deployment mechanism based on transfer learning is proposed to facilitate practical usage in the real satellite system, effectively promoting efficiency and thus saving precious onboard computational resources.

Author Contributions

Conceptualization, S.T., Z.P., G.H., Y.W. and Y.L.; Methodology, S.T., Z.P., G.H., Y.W. and Y.L.; Formal Analysis, S.T., Z.P., G.H. and Y.W.; Investigation, S.T. and Y.W.; Writing—Original Draft Preparation, S.T. and Z.P.; Writing—Review and Editing, S.T. and Z.P.; Visualization, S.T., Y.W. and Y.L.; Supervision, S.T., Z.P., G.H., Y.W. and Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This study was funded by National Natural Science Foundation of China (grant no. 62076251).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kua, J.; Loke, S.W.; Arora, C.; Fernando, N.; Ranaweera, C. Internet of Things in Space: A Review of Opportunities and Challenges from Satellite-Aided Computing to Digitally-Enhanced Space Living. Sensors 2021, 21, 8117. [Google Scholar] [CrossRef] [PubMed]
  2. Xu, Y.; Gui, G.; Gacanin, H.; Adachi, F. A survey on resource allocation for 5G heterogeneous networks: Current research, future trends and challenges. IEEE Commun. Surv. Tutor. 2021, 23, 668–695. [Google Scholar] [CrossRef]
  3. Munari, A.; Clazzer, F. Spectral Coexistence of QoS-Constrained and IoT Traffic in Satellite Systems. Sensors 2021, 21, 4630. [Google Scholar] [CrossRef] [PubMed]
  4. Isca, A.; Alagha, N.; Andreotti, R.; Andrenacci, M. Recent Advances in Design and Implementation of Satellite Gateways for Massive Uncoordinated Access Networks. Sensors 2022, 22, 565. [Google Scholar] [CrossRef] [PubMed]
  5. Qian, Y.; Wu, J.; Wang, R.; Zhu, F.; Zhang, W. Survey on reinforcement learning applications in communication networks. J. Commun. Inf. Netw. 2019, 4, 30–39. [Google Scholar]
  6. Luong, N.C.; Hoang, D.T.; Gong, S.; Niyato, D.; Wang, P.; Liang, Y.-C.; Kim, D.I. Applications of deep reinforcement learning in communications and networking: A survey. IEEE Commun. Surv. Tutor. 2019, 21, 3133–3174. [Google Scholar] [CrossRef] [Green Version]
  7. Chien, W.C.; Lai, C.F.; Hossain, M.S.; Muhammad, G. Heterogeneous space and terrestrial integrated networks for IoT: Architecture and challenges. IEEE Netw. 2019, 33, 15–21. [Google Scholar] [CrossRef]
  8. De Sanctis, M.; Cianca, E.; Araniti, G.; Bisio, I.; Prasad, R. Satellite communications supporting internet of remote things. IEEE Internet Things J. 2015, 3, 113–123. [Google Scholar] [CrossRef]
  9. Zhao, B.; Liu, J.; Wei, Z.; You, I. A deep reinforcement learning based approach for energy-efficient channel allocation in satellite internet of things. IEEE Access 2020, 8, 62197–62206. [Google Scholar] [CrossRef]
  10. Li, Z.; Wang, Y.; Liu, M.; Sun, R.; Chen, Y.; Yuan, J.; Li, J. Energy efficient resource allocation for UAV-assisted space-air-ground Internet of remote things networks. IEEE Access 2019, 7, 145348–145362. [Google Scholar] [CrossRef]
  11. Jia, M.; Zhang, X.; Gu, X.; Guo, Q.; Li, Y.; Lin, P. Interbeam interference constrained resource allocation for shared spectrum multibeam satellite communication systems. IEEE Internet Things J. 2018, 6, 6052–6059. [Google Scholar] [CrossRef]
  12. Liu, X.; Zhai, X.B.; Lu, W.; Wu, C. QoS-guarantee resource allocation for multibeam satellite industrial internet of things with NOMA. IEEE Trans. Ind. Inform. 2019, 17, 2052–2061. [Google Scholar] [CrossRef]
  13. Yan, X.; An, K.; Zhang, Q.; Zheng, G.; Chatzinotas, S.; Han, J. Delay constrained resource allocation for NOMA enabled satellite internet of things with deep reinforcement learning. IEEE Internet Things J. 2020, 20, 1–11. [Google Scholar]
  14. Zhang, X.; Zhang, B.; Guo, D.; An, K.; Qi, S.; Wu, G. Potential Game-Based Radio Resource Allocation in Uplink Multibeam Satellite IoT Networks. IEEE Trans. Aerosp. Electron. Syst. 2021, 57, 4269–4279. [Google Scholar] [CrossRef]
  15. Zhang, Y.; Zhang, H.; Zhou, H.; Long, K.; Karagiannidis, G.K. Resource Allocation in Terrestrial-Satellite-Based Next Generation Multiple Access Networks With Interference Cooperation. IEEE J. Sel. Areas Commun. 2022, 40, 1210–1221. [Google Scholar] [CrossRef]
  16. Sun, Y.; Wang, Y.; Jiao, J.; Wu, S.; Zhang, Q. Deep learning-based long-term power allocation scheme for NOMA downlink system in S-IoT. IEEE Access 2019, 7, 86288–86296. [Google Scholar] [CrossRef]
  17. Hu, X.; Liu, S.; Chen, R.; Wang, W.; Wang, C. A deep reinforcement learning-based framework for dynamic resource allocation in multibeam satellite systems. IEEE Commun. Lett. 2018, 22, 1612–1615. [Google Scholar] [CrossRef]
  18. Liu, S.; Hu, X.; Wang, W. Deep reinforcement learning based dynamic channel allocation algorithm in multibeam satellite systems. IEEE Access 2018, 6, 15733–15742. [Google Scholar] [CrossRef]
  19. Hu, X.; Liao, X.; Liu, Z.; Liu, S.; Ding, X.; Helaoui, M.; Wang, W.; Ghannouchi, F.M. Multi-agent deep reinforcement learning-based flexible satellite payload for mobile terminals. IEEE Trans. Veh. Technol. 2020, 69, 9849–9865. [Google Scholar] [CrossRef]
  20. Zhang, P.; Wang, X.; Ma, Z.; Liu, S.; Song, J. An online power allocation algorithm based on deep reinforcement learning in multibeam satellite systems. Int. J. Satell. Commun. Netw. 2020, 38, 450–461. [Google Scholar] [CrossRef]
  21. Shah, H.A.; Zhao, L.; Kim, I.M. Joint network control and resource allocation for space-terrestrial integrated network through hierarchal deep actor-critic reinforcement learning. IEEE Trans. Veh. Technol. 2021, 70, 4943–4954. [Google Scholar] [CrossRef]
  22. Xie, Y.; Ren, P.; Wang, Y.; Xu, D.; Li, Q.; Du, Q. Power-efficient uplink resource allocation for ultra-reliable and low-latency communication. In Proceedings of the IEEE 90th Vehicular Technology Conference (VTC2019-Fall), Honolulu, HI, USA, 22–25 September 2019; pp. 1–5. [Google Scholar]
  23. Watkins, C.J.C.H. Learning from Delayed Rewards. Ph.D. Thesis, King’s College London, London, UK, 1989. [Google Scholar]
  24. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.; Ostrovski, G.; et al. Human-level control through deep reinforcement learning. Nature 2015, 518, 529–533. [Google Scholar] [CrossRef] [PubMed]
  25. Paris, A.; Del Portillo, I.; Cameron, B.; Crawley, W. A genetic algorithm for joint power and bandwidth allocation in multibeam satellite systems. In Proceedings of the 2019 IEEE Aerospace Conference, Big Sky, MT, USA, 2–9 March 2019; pp. 1–15. [Google Scholar]
  26. Rui, W.; Xiao-dong, H.; Wei-yu, A.; Ke-zhen, S.; Huan, H. Research of Improved Genetic Algorithm for Resource Allocation in Space-based Information Network. In Proceedings of the International Conference on Wireless and Satellite Systems, Nanjing, China, 17–18 September 2020; Springer: Cham, Switzerland, 2020; pp. 139–152. [Google Scholar]
Figure 1. S-IoT scenario based on multibeam LEO satellite constellation.
Figure 1. S-IoT scenario based on multibeam LEO satellite constellation.
Sensors 22 02979 g001
Figure 2. Framework of DRL_CAPC algorithm.
Figure 2. Framework of DRL_CAPC algorithm.
Sensors 22 02979 g002
Figure 3. Convergence process of DRL-QoS-RA method.
Figure 3. Convergence process of DRL-QoS-RA method.
Sensors 22 02979 g003
Figure 4. The simulation process of each methods’ training process.
Figure 4. The simulation process of each methods’ training process.
Sensors 22 02979 g004
Figure 5. Deployment training process based on transfer learning.
Figure 5. Deployment training process based on transfer learning.
Sensors 22 02979 g005
Figure 6. Methods’ performance with requested traffic increase. (a) The trend of transmission success rate with traffic increase. (b) The trend of energy efficiency with traffic increase.
Figure 6. Methods’ performance with requested traffic increase. (a) The trend of transmission success rate with traffic increase. (b) The trend of energy efficiency with traffic increase.
Sensors 22 02979 g006
Table 1. Detailed parameters of S-IoT scenario.
Table 1. Detailed parameters of S-IoT scenario.
ParametersValue
Satellite altitude550 km
Beams19
Data transmission channels8
Channel bandwidth10 KHz
Frequency band14 GHz
Terminals’ antenna max power300 mW
Terminals’ antenna power level3
Terminals’ antenna gain10 dBi
Satellite receiving antenna G/T3.7 db
Path loss170.38 db
Satellite backhaul power limitation300 W
Beam power limitation25 W
Satellite amplifier magnification5
SINR threshold1.1
Table 2. DRL-QoS-RA algorithm parameters.
Table 2. DRL-QoS-RA algorithm parameters.
Algorithm ParametersValue
Replay start size2000
Replay memory20,000
Batchsize32
Target network update step50
Discount factor0.99
Initial exploration rate1.0
Final exploration rate0.01
Exploration rate decay5 × 10−4
Learning rate0.001
Table 3. Performance of comparative methods with different request arrival rates.
Table 3. Performance of comparative methods with different request arrival rates.
Methodsλ = 1.5λ = 9.75
Success RateEnergy EfficiencyComputational Time (s)Success RateEnergy EfficiencyComputational Time (s)
random0.940.16-0.340.12-
GA0.990.413.24 × 1030.530.221.67 × 104
DRL-RA10.2764.770.890.2364.77
DRL-EERA10.6971.520.480.3871.52
DRL-QoS-RA10.6678.310.710.3378.31
DRL-QoS-RA-transferred10.6411.530.720.3111.53
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tang, S.; Pan, Z.; Hu, G.; Wu, Y.; Li, Y. Deep Reinforcement Learning-Based Resource Allocation for Satellite Internet of Things with Diverse QoS Guarantee. Sensors 2022, 22, 2979. https://doi.org/10.3390/s22082979

AMA Style

Tang S, Pan Z, Hu G, Wu Y, Li Y. Deep Reinforcement Learning-Based Resource Allocation for Satellite Internet of Things with Diverse QoS Guarantee. Sensors. 2022; 22(8):2979. https://doi.org/10.3390/s22082979

Chicago/Turabian Style

Tang, Siqi, Zhisong Pan, Guyu Hu, Yang Wu, and Yunbo Li. 2022. "Deep Reinforcement Learning-Based Resource Allocation for Satellite Internet of Things with Diverse QoS Guarantee" Sensors 22, no. 8: 2979. https://doi.org/10.3390/s22082979

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop