Abstract
This paper proposed an adaptive three-dimensional (3D) path-following control design for a robotic airship based on reinforcement learning. The airship 3D path-following control is decomposed into the altitude control and the planar path-following control, and the Markov decision process (MDP) models of the control problems are established, in which the scale of the state space is reduced by parameter simplification and coordinate transformation. To ensure the control adaptability without dependence on an accurate airship dynamic model, a -Learning algorithm is directly adopted for learning the action policy of actuator commands, and the controller is trained online based on actual motion. A cerebellar model articulation controller (CMAC) neural network is employed for experience generalization to accelerate the training process. Simulation results demonstrate that the proposed controllers can achieve comparable performance to the well-tuned proportion integral differential (PID) controllers and have a more intelligent decision-making ability.
1. Introduction
The lift of an airship mainly comes from the buoyancy of the gas, and therefore, it does not have to do continuous movement to balance gravity, which makes it a promising platform for long-endurance and low-energy consumption. In the last two decades, airships such as heavy-duty airships and stratospheric airships [1] have gained increasing interest given their potential in communication relay, space observation, and military reconnaissance. Control strategy has always been an important subject in airship research. After years of study, airship motion control has gained significant achievements with regard to hovering control, trajectory tracking, and path-following control [2–4]. The objective of the hovering control is to maintain the airship in a certain area against the wind, and the primary purpose of the trajectory tracking is to track a time-parameterized desired path, while the path-following control is mainly concerned with tracking a desired path without a specified temporal constraint.
The control response of an airship is slow with a long time delay. The flexibility and elasticity of the hull and the stochastic atmosphere disturbance make the control of the airship a problem with uncertainties. At present, common airship control approaches include PID control, backstepping, dynamic inversion, sliding mode control, and robust control [5–8]. The PID controller is widely used for its simplicity and effectiveness, but its parameters such as proportion, integral, and differential should be tuned properly [9]. When the model of the system or environment changes, the performance of the PID controller may get worse before being tuned again. The modern model-based nonlinear control approaches can ensure the control robustness and global stability while being accurately controlled; however, these improvements incur the work of system modeling and parameter identification with ever increasing complexity [10, 11]. With the rapid development of information technology, machine learning algorithms have been successfully applied in various complex applications [12] and can make intelligent decisions similar to human beings. As an important unsupervised learning algorithm, reinforcement learning is able to realize adaptive control and decision in an unknown environment through the mechanism of “action and reward” [13, 14], which does not need an accurate model that costs human effort.
Up to now, reinforcement learning has been extensively studied in robot control, traffic dispatch, communication control, and game decision-making [15–18]. In the field of aircraft navigation and control, reinforcement learning also witnessed many successful applications. However, due to the complexity of aircraft motion and the large scale of the state space, the reinforcement learning control can easily be trapped in the “curse of dimensionality,” which confines the applications to outer-level decisions rather than inner actuator commands. In particular, Pearre and Brown [19] and Dunn et al. [20] took the flight path angle as the action of the reinforcement learning algorithm and constructed the reward function using energy consumption, time spent, and collision loss to flexibly plan and optimize the reference flight path. To further optimize the flight path, Zhang et al. [21] proposed a “geometric reinforcement learning” algorithm to provide more available actions by transferring between nonadjacent states. Palunko et al. [22] and Faust [23] introduce a reinforcement learning algorithm to the adaptive acceleration control of a quadrotor with suspended load. The primary purpose of the above studies is to generate the outer loop guidance law; however, the inner loop controllers still need to be designed. In [24], a reinforcement learning approach for the airship model parameter identification is suggested by Ko et al., which can reduce the control error caused by the model uncertainties. In a previous work in [25], Rottmann et al. designed an airship altitude controller based on reinforcement learning and completed the online training process in a few minutes. However, the controller training of [24, 25] is performed in the restricted low-dimensional state space, and therefore, the approach is difficult to be directly employed in the multidimensional motion control. Hwangbo et al. [26] proposed a neural network quadrotor control strategy based on a reinforcement learning algorithm, and the well-trained controller can achieve autonomous flight control even if the quadrotor is thrown upside down into the air, which is superior to human operators. But in [26], the quadrotor’s model was used for training the neural network controller off-line to avoid the “curse of dimensionality” problem, which brought in the risk of control failure when the dynamic model is inaccurate or unavailable.
For the purpose of acquiring a feasible airship control strategy without dependence on an accurate dynamic model, this paper directly takes the inner actuator commands as the action of the reinforcement learning algorithm and implements the training by actual motion. The main contributions of this paper are as follows: (1)The MDP models of the airship 3D path-following control are established based on motion analysis, which provide new models and methods for airship control. Furthermore, in contrast to the previous reinforcement learning airship control without dependence on an accurate dynamic model, for the first time, this study proposed an airship control strategy through autonomous online training in 3D space(2)To deal with the “curse of dimensionality” problem in the airship planar path-following control, a novel coordinate frame is proposed to describe the relative state between the airship and the target. This coordinate transformation form can reduce the scale of the state space and generalize the experience by rotation and makes it possible to learn the control strategy online(3)In the proposed reinforcement learning airship control strategy, a CMAC neural network is employed to generalize the experience in a local neighborhood, which effectively accelerates the training process
The rest of the paper is organized as follows. Section 2 presents the problem formulation and establishes the MDP models of the airship control. Section 3 introduces the structure, the principles, and the training process of the proposed airship reinforcement learning control strategy. Section 4 presents the simulation results, and Section 5 concludes the paper.
2. Problem Formulation and the MDP Model of the Airship Control
2.1. Problem Formulation
After takeoff and ascent, the airship stays steadily near its design altitude. Thereafter, the altitude of the airship is mainly controlled by an elevator, and the horizontal motion is mainly controlled by a propeller, rudder, and vector propeller. Compared with the fixed-wing aircraft, the airship has a smaller roll angle, and the rolling motion has little effect on the horizontal motion; hence, the coupling of the airship longitudinal motion and lateral motion is weak [27]. To simplify the control problem, we decompose the airship 3D path-following control into the altitude control and the planar path-following control.
The autonomous training process of reinforcement learning is similar to the animal learning process, in which an intuitive objective will contribute to the acceleration of learning. In this paper, the airship completes the path-following mission by reaching the target points on the desired path one by one. represents the earth reference frame (ERF), where denotes the position of the airship. Let denote the altitude of the airship which satisfies ; then, the position of the airship can be described as .
In the airship altitude control, suppose the desired altitude is and the control objective is given by where is defined as the valid range of the desired altitude.
In the airship planar path-following control, let denote the position of the target point, and the control objective can be described as where indicates the valid area of the target point.
Remark 1. During the training process, the desired altitude and the target point of the two controllers can be set randomly and independently. In the 3D path-following control, we choose the planar path-following control as the main control mission and the altitude control as the auxiliary. To ensure that the actual path is close to the desired 3D path, the horizontal position of the current target is taken as the target point in the planar path-following control and the altitude of the nearest point on the desired path to the airship is taken as the desired altitude in the altitude control. Let denote the current target, denote the last reached target, and denote the position of the airship. Then, the target point in the planar path-following control can be set as and the desired altitude in the altitude control can be calculated as where and are given by
2.2. MDP Model of the Airship Control
MDP is the basis for the implementation of reinforcement learning; thus, we establish the MDP models of the airship altitude control and airship planar path-following control first. MDP can be expressed by five elements , where indicates the state space constructed by parameters such as the position, speed, and attitude of the airship; is the action set composed of available control actuator commands, is the reward of the state and action, is the state transition probability, and is the optimization objective function of the sequential decision. MDP satisfies the following property: where is defined as the transition probability from state to state after taking the action , and is the time parameter. It can be seen from the property (5) that parameters of the state space are sufficient statistics of the airship motion. The primary work of airship control MDP modelling is to construct the state space. By reasonably selecting parameters to construct state space, the property (5) can be approximately satisfied.
2.2.1. State Space of the Altitude Control
As shown in Figure 1, let denote the earth reference frame (ERF) and denote the body reference frame (BRF) attached to the airship. The attitude of the airship is described as the Euler angles in EBF. The speed and angular velocity of the airship in BRF are defined as and , respectively. , , and represent the rudder deflection, the elevator deflection and the propeller thrust of the airship, respectively.

The airship resultant velocity is defined as in ERF, which satisfies the following equation:
According to the kinematics of the airship, the speed in ERF can be calculated as
is the rotation matrix from BRF to ERF, which is given by where is and is .
According to (7) and , the altitude-changing rate can be calculated as
In the actual motion of the airship, the speeds and are much smaller than , and due to the strong restoring torque and large damping of the rolling motion, the roll angle is small. Therefore, from (6) and (9), approximate equations can be obtained as
It is can be seen that the three parameters of the velocity , the pitch angle , and the altitude-changing rate have strong correlation.
The airship motion has 12 state parameters of position, velocity, angle, and angular velocity in 3 coordinate axes. If all the parameters are taken into account in the airship altitude control, the scale of the state space and the amount of memory to store value function estimates will be too large for the controller to complete the training in a limited time. To avoid this “curse of dimensionality” problem, it is necessary to simplify the state parameters according to their relevance to motion control.
The parameters related to the airship altitude control mainly include the altitude , the altitude-changing rate , the velocity , and the pitch angle . From (10), , , and have strong correlation; thus, two of them can contain the main information. Since the speed parameters are relatively easy to measure, we construct the state space of the airship altitude control MDP model by .
2.2.2. State Space of the Planar Path-Following Control
According to the control objective (2), the state parameters related to the planar path-following control mainly include the position of the airship (, ), the position of the target (, ), the velocity , and the course angle . We use the parameters to describe the main information of the airship planar path-following control.
Remark 2. The yaw angle of airship is also an important parameter for the airship path-following control. However, the course angle has strong correlation with the yaw angle due to the course stability of the airship. Under the influence of the airship vertical tail, the course angle will soon follow the yaw angle , and the difference between them is limited. Since the course angle is directly related to the planar motion, we choose to describe the state of the airship.
To avoid the “curse of dimensionality” problem during the training process in multidimensional space, the parameters can be simplified and merged based on airship motion analysis. A common method is to rebuild the coordinate frame with a relative position. Denote target position as the origin, and the relative position can be calculated as and ; then, the state of the airship can be expressed as . To further accelerate the training process of the planar path-following controller, the experience can be generalized by rotating around the origin. Inspired by this idea, we present a novel coordinate frame to describe the relative state between the airship and target, as shown in Figure 2.

In Figure 2, represents the distance from the airship to the target, is the angle between the axis and the line from the airship to the target, and is the angle between the line from the airship to the target and the direction of . According to the geometrical relation, the distance , the angle , and the velocity can describe the relative state between the airship and target, which are calculated by
By means of coordinate transformation, the state space of the planar path-following control can be constructed by , and the dimensions of the planar path-following control state space is reduced from 6 to 3, which effectively avoids the “curse of dimensionality” problem in multidimensional online training.
2.2.3. of the Airship Control
The action set of MDP can be flexibly adjusted according to the actual control actuators. In this study, the action set of the airship altitude control is constructed by elevator deflection and of the planar path-following control is constructed by rudder deflection. The reward function consists of the positive reward wherein the airship reaches the valid target area and the negative reward wherein the airship flies outside the border. In this study, the controller is trained online and the state transition probability is determined by the actual response of the airship to the control output. The optimization objective function is set as the total cumulative reward. Let denote the action policy and denote the optimal cumulative reward, which is given by where is the reward at time and is the reward discount factor.
3. Path-Following Control Design
The reinforcement learning altitude controller and the reinforcement learning planar path-following controller have similar structures as shown in Figure 3. The mapping between the “state input” and the “control output” is created autonomously through online training.

As can be seen from Figure 3, the airship reinforcement learning controller consists of three modules: the error calculation module, the CMAC neural network module, and the action policy module. The functions of each module are as follows.
3.1. Error Calculation
The error of the value function estimation is calculated based on the reinforcement learning algorithm according to the actual state transition of the airship. This error signal will be used for updating the weights of the CMAC neural network.
3.2. CMAC Neural Network
The value function of each action is estimated by the CMAC neural network, which provides the decision-making basis for the action policy module.
3.3. Action Policy
The main function of this module is to balance between exploration and the greedy policy according to the value function estimations of actions and then output the corresponding control command.
Remark 3. Different from the supervised learning, the error for modifying the neural network in the reinforcement learning controller is not from the sample’s label but is calculated from the estimation of the value function and reward function. In fact, if there are enough training samples with labels provided by an expert controller, similar structures can be used for imitating the control strategy of the expert controller.
3.4. -Learning Algorithm
The reinforcement learning controller improves its action policy by interacting with the external environment. The controller selects the action based on the estimation of the value function and updates the estimation to maximize the cumulative reward which is given by the external environment with a random delay. Based on this concept, the -Learning algorithm estimates and stores the action value function in a different state and uses a “trial and error” approach to update the estimation of the action value function. -Learning is an appropriate reinforcement learning algorithm for the online training of the airship controller, and the properties of the airship also make it a suitable platform for the -Learning algorithm. Compared with an airplane and a helicopter, the attitude of the airship is relatively stable, and the “trial and error” operations of the airship rarely cause a catastrophic consequence. Furthermore, the airship is a platform with long endurance, which can provide enough time for the training and iterative function estimation of the algorithm. In this paper, the -Learning algorithm is adopted for calculating the modifying error of the neural network.
Let denote the value function estimation of the action in state , which can be calculated as where is the reward at time , is the reward discount factor, is the initial state, and is the initial action.
According to the theory of operation research, satisfies the following Bellman equation: where denotes the transition probability from state to state after taking the action and indicates the reward of the action and the transition from to . Let denote the action policy; then, the optimal action policy and the optimal action estimation corresponding to equation (14) can be described as follows:
From (14) and (15), the one-step update formula of under the optimal action policy can be obtained as where is the learning rate of the -Learning algorithm.
3.5. CMAC Neural Network
In actual control applications, the multidegrees of system freedom always result in the large scale of the state space. In order to accelerate the training process of the reinforcement learning controller, an effective approach is to fit the value function based on the theory of stochastic process and the neural network. Neural networks can be divided into global approximation neural networks and local approximation neural networks. For global approximation neural networks like the Back Propagation (BP) neural network, each weight may affect the output and each training sample will update all the weights. As a result, the convergence speed is slow and can hardly meet the requirement of online training. For local approximation neural networks like the CMAC neural network, only a few weights affect the output and each training sample only updates the local weights, which is able to quickly approximate time-varying nonlinear functions.
During the training of the reinforcement learning controller, there are not enough training samples but only limited data from online motion; hence, a rapid convergence speed is necessary. For an airship, the control strategies are alike in similar states, and the control strategies may be completely different in very dissimilar states. The advantages of the local approximation neural network suit these properties of the airship. The CMAC neural network fits the nonlinear function by table querying and updates the weights in the local neighborhood, and it has less computation, which is convenient to be applied in the embedded system of a flight control computer. For the above advantages, in this study, the CMAC neural network is employed to generalize the training experience of the airship reinforcement learning controller.
The structure of a typical CMAC neural network is shown in Figure 4, where is the state space formed by states, is the weights stored in memory addresses, and is the output of the network. The actual state input is mapped to state in the state space , and each state corresponds to an activation vector . The and are given by where represents the activation state of to . For any two states, the closer the states are, the more overlaps their activation vectors have. And if two states are far away from each other, their activation vectors have no overlap.

The output of the CMAC neural network can be obtained as
We establish one CMAC neural network for each action in the action set , and the network weights of can be defined as . Letting denote the value function estimation of the state-action pair , we have
Then, the partial derivative of to is
Suppose as the action at time , then in state , the value function estimation error of is given by
The network weights of action can be updated by a gradient algorithm where is the learning rate of the weights. By substituting (20) and (21) into (22), we have but is unknown. Based on the concept (16) of the -Learning algorithm, can be replaced by which is calculated from the reward and value function estimation at time . Taking into account the impact of the state visit frequency, the updated formula of with a fitness coefficient can be written as where is the fitness coefficient with an initial value of 1, which reflects the state visit frequency:
3.6. Process of Training and Control
To avoid being trapped in a local optimum and ensure the convergence speed of the training, a proper action policy is necessary for the reinforcement learning controller. In this paper, a random action policy based on the Boltzmann distribution is adopted for balancing between the exploration and greedy policy. Let denote the available action set and denote the probability of selecting action in state ; then, the action policy is given by where is the exploration coefficient which is similar to the temperature coefficient of a simulated annealing algorithm. By progressively reducing , this policy ensures a high exploration probability in the initial states and gradually trends to the greedy policy.
The training and control process of the proposed airship reinforcement learning controller is described as shown in Figure 5. The flowchart before the dotted line is the online training process, and the rest is the actual control process.

As can be seen from Figure 5, during the training process, the airship starts one round of learning after another by randomly setting the position of the target. In each round of learning, the controller selects the action based on policy periodically before reaching the target or boundary, and the fitness coefficient and the weights of the CMAC neural network are updated at the same time. When the algorithm and the neural network converge, the training process is completed and the controller can be applied in actual control missions. During the control process, the parameters of the algorithm are fixed, and the action policy is turned into the greedy policy. The controller maximizes the cumulative reward by selecting the action corresponding to the maximum value function estimation.
4. Simulation
To test the performance of the proposed controllers, simulations are implemented in MATLAB. We build the model of a LS-S1200 robotic airship [28] based on the approach in [9, 27]. The aerodynamic forces and torque can be calculated as where is the dynamic pressure of the inflow, is the atmospheric density, is the attack angle, is the sideslip angle, are the aerodynamic forces and torque of the airship, and are the hull aerodynamic coefficients. The aerodynamic coefficients, the structure parameters, and the reinforcement learning controller parameters used in the simulation are listed in Table 1, where is the mass of the airship, is the volume of the airship, are the moments of inertia, and are the inertia factors.
During the training of the controller, the propeller thrust changes randomly every 15 seconds to simulate the variation of the airship velocity. When the training is completed and the control mission begins, the airship velocity is maintained by a simple proportion integral (PI) propeller thrust controller, in which the cruising velocity is set to 10 m/s.
For convenience in analysing the control performance, the well-tuned PID controller is chosen as a comparison. In the outer loop of the PID controller, the desired attitude angle is calculated based on the line-of-sight (LOS) guidance law [29]. In the inner loop, we establish the dynamic equations according to the airship’s structure and aerodynamic parameters and then linearize the equations with a small perturbation method. After working out the transfer function from the rudder deflection to the attitude angle, the gain parameters of the PID controller can be tuned well using the root locus tool RLTOOL in MATLAB.
4.1. Training of Reinforcement Learning Controllers
The training of the proposed airship altitude control and planar path-following control is independent and can be implemented simultaneously. The controller is trained online by randomly setting the target near the current position of the airship. The setting range of the desired altitude is 10 m, and the setting range of the target point is 100 m. In this paper, the weights of the CMAC neural network are saved every 100 s for test and evaluation. Figure 6 shows the trend of the success rate and the average time spent on the training missions during the training process.

(a) Success rate of missions in the altitude controller training

(b) Average time spent of missions in the altitude controller training

(c) Success rate of missions in the planar path-following controller training

(d) Average time spent of missions in the planar path-following controller training
In Figure 6(a), after training for about 1000 s, the mission success rate of the airship altitude controller is stable at 100%; then, the altitude controller can be applied in actual control missions. And as shown in Figures 6(c) and 6(d), the training of the airship planar path-following controller is completed in about 3 hours. In Figures 6(b) and 6(d), the average time spent on the training missions increased first, then decreased gradually, and eventually tended to be stable. This process corresponds to three stages of learning: at the initial stage, the controller could not control the airship and failed soon; then, the controller learned to lead the airship to reach the target in an indirect and suboptimal path; finally, the action policy was continuously optimized and gradually stabilized.
4.2. Altitude Control
We use the trained reinforcement learning controller and a well-tuned PID controller to implement the airship altitude control mission. When the propeller thrust remains constant and the airship keeps a straight flight (), the simulation result is shown in Figure 7(a). When the propeller thrust and rudder deflection change randomly every 15 seconds, the simulation result is shown in Figure 7(b).

(a) Altitude control with constant and

(b) Altitude control with varied and
As shown in Figure 7(a), both the reinforcement learning controller and the PID controller perform well when the airship is in the steady flight state. The altitude control errors of the two controllers are within 5 m, and the PID controller performs slightly better.
In Figure 7(b), it can be seen that when the airship is in the maneuver flight state, the performance of the reinforcement learning controller is slightly affected, and its altitude control error is still within 5 m. But the PID controller is significantly affected and its maximum altitude control error reaches about 10 m, which is consistent with the actual experiment data of the airship [28]. The reinforcement learning controller performs better than the PID controller when the airship is in the maneuver flight state. This is because during the training process, the airship has been in a random maneuver flight state, and the reinforcement learning controller gradually learned the action policy in that state by autonomous adjustment.
4.3. Planar Path-Following Control
A desired path consisting of continuous turns, a straight line, and arcs is used for testing the performance of the planar path-following controller. Figure 8 shows the planar path-following control simulation results. Figure 9 shows the deviation distance of the airship to the target points and the desired path.


(a) Distance from the airship to the target points

(b) Distance from the airship to the desired path
In Figures 8 and 9, it can be seen that the planar path-following control error of the reinforcement learning controller is close to that of the PID controller near the 20th target point. At this point, the airship is in the steady flight state on the desired straight path. During the rest of the flight, the airship is in the maneuver flight state, and the control error of the reinforcement learning controller is smaller. This is because the action policy of the reinforcement learning controller is not determined by the current deviation but based on the total cumulative reward. When the current deviation is small but a significant control error is about to occur, the reinforcement learning controller can react in time.
The deviation distance of the reinforcement learning controller and the PID controller is compared in Table 2. To complete the planar path-following control mission, the reinforcement learning controller takes 419.2 s and the PID controller takes 443.7 s. By comparing the above data, it is clear that the reinforcement learning controller not only is slightly better in terms of control accuracy but also takes less time in the whole control mission.
To test the adaptability of the reinforcement learning controller, we can assume that asymmetric aerodynamic shape changes occur in the airship dynamic model. In this example, an additional yaw torque coefficient () is added to the dynamic equation, which will lead to the asymmetric changes in the airship’s turning radius and turning ability. As a result, the performance of the original controllers may get worse, but the reinforcement learning controller can autonomously adjust its action policy by training online. Figure 10 shows the simulation result of the controllers after the airship’s dynamic model changed.

As shown in Figure 10, the airship’s right turning ability is weaker than the left due to the asymmetric aerodynamic shape. Controlled by the PID controller, the airship kept circling around the 2nd target point but cannot get closer. Through expanding the valid target area, the PID controller can complete the control mission, but its control error becomes larger because of the relaxation of the control requirements. After online training, the reinforcement learning controller autonomously adapts to the model changes and is able to complete the control mission without expanding the valid target area. Furthermore, it is remarkable that near the 11th target point, the reinforcement learning controller shows an intelligent decision-making ability similar to human operators. When the airship could not reach the 11th target point directly because of the weak right turning ability, the reinforcement learning controller implemented the action policy of flying further at first and then turning left, which successfully led the airship to reach the target. This is owing to the experience of reaching the target in an indirect path that the reinforcement learning controller gained during the training process.
4.4. 3D Path-Following Control
The 3D path-following control mission can be completed through the cooperation of the proposed controllers. In this paper, the target point of the planar path-following controller and the desired altitude of the altitude controller are calculated according to the following method.
Planar path-following control: in the 3D path-following control, the planar path-following is chosen as the main control mission; therefore, we project the 3D desired path onto the horizontal plane and create a sequence of target points.
Altitude control: while the airship is flying to the target point, the altitude controller works as an auxiliary and updates its desired altitude periodically. As shown in Remark 1, the altitude of the nearest point on the 3D desired path to the airship is taken as the desired altitude.
Taking a spiral path as an example, the simulation result of the proposed airship reinforcement learning controller is shown in Figure 11, where the horizontal distance between two adjacent target points is 100 m and the vertical distance is 10 m. It can be seen from Figure 11 that the performance of the planar path-following controller in the 3D path-following control is similar to that in the independent planar control, while the altitude of the airship slightly shows step changes. This is because the desired altitude of the altitude controller is updated periodically. The overall simulation result shows that the reinforcement learning planar path-following controller and altitude controller do not interfere with each other and cooperate well.

5. Conclusions
In this paper, a reinforcement learning airship 3D path-following control is proposed, which is robust against dynamic model changes. To avoid the “curse of dimensionality” problem, the MDP models of the airship altitude control and planar path-following control are established, and the scale of the state space is reduced by parameter simplification and coordinate transformation. The airship reinforcement learning controllers are designed based on the -Learning algorithm and CMAC neural network, and the training process can converge in 3 hours. The reinforcement learning altitude controller and planar path-following controller can achieve comparable performance to well-tuned PID controllers, and the 3D path-following control mission can be completed through cooperation of the proposed controllers. Compared with existing approaches, the main advantage of this control strategy is that when the airship model dynamic parameters are uncertain or wrong, the controller is able to adjust its action policy autonomously and make intelligent decisions similar to human operators.
Data Availability
The data used to support the findings of this study are available from the corresponding author upon request.
Conflicts of Interest
The authors declare that there is no conflict of interest regarding the publication of this paper.
Acknowledgments
This work was supported by the National Natural Science Foundation of China (No. 61503010), the Aeronautical Science Foundation of China (No. 2016ZA51001), and the Fundamental Research Funds for the Central Universities (No. YWF-19-BJ-J-118).