Next Article in Journal
Novel Meta-Fractal Wearable Sensors and Antennas for Medical, Communication, 5G, and IoT Applications
Next Article in Special Issue
Robust Fractional-Order PI/PD Controllers for a Cascade Control Structure of Servo Systems
Previous Article in Journal
Understanding the Complex Adaptive Characteristics of Cross-Regional Emergency Collaboration in China: A Stochastic Evolutionary Game Approach
Previous Article in Special Issue
A Fractional-Order ADRC Architecture for a PMSM Position Servo System with Improved Disturbance Rejection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fractional-Order Control Method Based on Twin-Delayed Deep Deterministic Policy Gradient Algorithm

1
College of Automation Engineering, Nanjing University of Aeronautics and Astronautics (NUAA), Nanjing 211106, China
2
Jincheng Nanjing Engineering Institute of Aircraft Systems, Nanjing 211199, China
*
Author to whom correspondence should be addressed.
Fractal Fract. 2024, 8(2), 99; https://doi.org/10.3390/fractalfract8020099
Submission received: 30 December 2023 / Revised: 23 January 2024 / Accepted: 3 February 2024 / Published: 6 February 2024
(This article belongs to the Special Issue Applications of Fractional-Order Systems to Automatic Control)

Abstract

:
In this paper, a fractional-order control method based on the twin-delayed deep deterministic policy gradient (TD3) algorithm in reinforcement learning is proposed. A fractional-order disturbance observer is designed to estimate the disturbances, and the radial basis function network is selected to approximate system uncertainties in the system. Then, a fractional-order sliding-mode controller is constructed to control the system, and the parameters of the controller are tuned using the TD3 algorithm, which can optimize the control effect. The results show that the fractional-order control method based on the TD3 algorithm can not only improve the closed-loop system performance under different operating conditions but also enhance the signal tracking capability.

1. Introduction

Fractional-order calculus stands out for its flexible description of the behavior of nonlocal and non-Markovian dynamics, providing a rich mathematical tool for modeling complex systems. Its main applications include the modeling of nonlinear and nonsmooth system dynamics [1], the simulation of multiscale complex systems [2], the analysis of non-Markovian processes [3], as well as in the fields of signal processing, control systems and financial modeling [4]. In the field of control, fractional-order control has been combined with many traditional control schemes, such as fractional-order PID control [5], fractional-order robust control [6], and fractional-order sliding-mode control [7]; these fractional-order control methods have been developed in depth at the theoretical level to form a sound theoretical system and have achieved extensive and powerful results in practical applications.
As a traditional control strategy, fractional-order sliding-mode control is able to cope with the nonlinearity and uncertainty of the system more flexibly [8,9,10]. However, with the increase in system complexity and nonlinearity, choosing the optimal values of controller parameters often becomes a great difficulty [11]. Faced with this problem, different scholars have used a variety of optimization algorithms over the years, such as early adaptive control [12], the particle swarm algorithm [13], the genetic algorithm [14], and the wolf-pack algorithm [15]. These algorithms have had some success, but as technology advances, they struggle to handle increasingly complex problems.
In recent years, with the continuous development of artificial intelligence technology, more and more advanced algorithms have been applied to the control field [16]. Among them, reinforcement learning algorithms based on neural networks are increasingly becoming a focus of research because their flexibility and adaptability make them more adaptable to complex environments [17,18,19,20,21,22,23]. Reinforcement learning efficiently masters system dynamics by learning through the interaction of an agent with the environment [24]. This approach transforms the system control problem into a process in which the agent learns the optimal control strategy through continuous trial and error [25], and it can also learn without prior knowledge of the system model [26,27], which offers a flexible approach for real-time controller parameter optimization. Specifically, under the framework of reinforcement learning, the intelligent body adjusts the control strategy according to feedback signals by interacting with the environment and gradually optimizes the controller parameters [28,29]. This learning approach is more adaptive, especially in the face of a high-order system complexity, significant nonlinearities, and rapid changes in dynamic characteristics, where traditional static parameter optimization methods may appear to be inadequate [30]. Therefore, reinforcement learning provides a more intelligent solution for controller parameter optimization, which is expected to show more significant performance improvements in practical applications.
In this paper, a fractional-order control method based on TD3 reinforcement learning is proposed to optimize the parameters. A fractional-order disturbance observer is designed for estimating the disturbance signals present in the system, while an RBF network is designed for approximating the possible uncertainties in the system, and finally, a fractional-order sliding-mode controller is designed for controlling the system. For the parameter optimization part, it is performed by the TD3 reinforcement learning algorithm, which consists of six deep neural network networks designed to inhibit the bootstrap phenomenon in the reinforcement learning algorithm [31], so that the output of the network can converge to the optimal solution quickly. On this theoretical basis, a valve-controlled hydraulic system is selected for design and simulation in Matlab/Simulink to verify the effectiveness of the method proposed in this paper, and at the same time, in order to reflect the advantages and disadvantages of the proposed method, a series of comparative experiments are designed.
In summary, the main work of this paper can be succinctly summarized in the following three areas:
i.
The fractional-order disturbance observer is designed to estimate the system disturbance signal, and the RBF network is selected to approximate the uncertainties of the system; then, a fractional-order sliding-mode controller is designed to control the system according to the estimated value of the disturbance observer;
ii.
The TD3 algorithm is introduced to optimize the parameters of the controller, and an improved loss function is designed to improve the learning performance of the algorithm, accelerate the convergence of the network output, and optimize the control effect;
iii.
It is verified through simulation that not only the control effect of the proposed method is better than the selected comparison control method, but that its also has a great robustness and generalization capability.
This paper is organized as follows. Section 2 introduces the proposed fractional-order control method, including the system state equation generalization and the designed fractional-order disturbance observer with an RBF network structure, based on which the fractional-order sliding-mode controller is designed. Section 3 introduces the fundamentals of the TD3 algorithm [32] and defines the reward function and loss function used. Section 4 demonstrates the stability of the designed method to verify its theoretical correctness. In Section 5, the proposed method is simulated in Matlab/Simulink using a valve-controlled hydraulic system as a model, and a series of control simulations are performed to verify the practical feasibility of the proposed method. Finally, conclusions and future work are presented.

2. Design of Fractional-Order Control Method

In this section, the fractional-order control part of the proposed method is accomplished for a system generalization including a fractional-order disturbance observer, an RBF network, and a fractional-order sliding-mode controller.
For a common system, the equation of state can be expressed as
x ˙ = A x + B u + f x + d y = C x
where x is the system state, and x R n ; A is the system parameter matrix, A R n × n , B is the control matrix, and B R n × q ; u is the control signal, u R q , C R p × n is the output matrix, d R n × 1 is the disturbance signal, and f ( x ) R n × 1 is the uncertainty of system.
In this paper, the Riemann–Liouville fractional-order calculus formula is defined as [33]:
D α g x = 1 Γ n α d n x n a x g t x t α + 1 n d t
where α represents the differential, 0 < α < 1 , Γ · represents the Gamma function, n represents the smallest integer larger than α , usually taken as 1, and a, x represents the lower and upper limits of the integral.
For the disturbance signal, the fractional-order disturbance observer is given by
D λ z = L D λ 1 z + L D λ 1 x L A D λ 1 x + B D λ 1 u + D λ 1 f ^ ( x ) d ^ = z + L x
where z is the disturbance observer auxiliary vector, L is the gain matrix, L is the Hurwitz matrix with n distinct eigenvalues, d ^ is the estimate of the disturbance, and f ^ ( x ) is an estimate of the model uncertainty.
Theorem 1.
Assume that the disturbance is bounded and its derivative is also bounded, which means that d t ς , d ˙ t ζ . Define the estimation error of the disturbance e d t = d t d ^ t . Based on the structure of the designed disturbance observer, the estimation error of the disturbance is bounded when the system uncertainty estimation error f ˜ ( x ) is bounded, which means that e d t ξ , where ξ is a very small constant greater than zero.
Proof. 
Define M = L ; then, M is a Hurwitz matrix with n distinct eigenvalues, so there exists an invertible matrix X such that X 1 M X = d i a g λ 1 , λ 2 , , λ n . Therefore, there exists a positive constant σ such that e M t σ e λ max M t , where σ = X 1 X [34]. Differentiating the estimation error e d t for the disturbance yields the following equation: d e d t d t = d ˙ t d ^ ˙ t = d ˙ t d d t z + L x = d ˙ t + L z + L x + L A x + B u + f ^ ( x ) L A x + B u + f ( x ) + d = d ˙ t + L d ^ t L d t + L f ^ ( x ) L f ( x ) = d ˙ t L e d t L f ˜ ( x )
It can be obtained that e ˙ d t = d ˙ t + M e d t + M f ˜ ( x ) . The subsequent proof is given in the stability analysis on e d ( t ) .   □
For the model uncertainty that may exist in the system, a radial basis function (RBF) network is used for approximation. The network structure of an RBF has the ability for high-speed learning, the ability to approximate a nonlinear function, which improves the model’s fitting ability, and it can adapt to a variety of complex input–output mapping, so it is very flexible in practical applications. Moreover, compared with other types of neural networks, the RBF network structure is relatively simple, which makes it easy to realize and adjust [35]. The structure of the RBF network used in this paper is shown in Figure 1.
The input–output relationship for each layer can be expressed as an input layer, θ i = x i , i = 1 , 2 , , n , an implicit layer, h j = e n e t j , n e t j = j = 1 m θ j c j 2 b j 2 , j = 1 , 2 , m , and an output layer, Y i = W i T H = ω i 1 h 1 + ω i 2 h 2 + + ω i m h m , where · is the Euclidean paradigm, H is an implicit layer function vector H = h 1 h 2 h m T , c j is the center vector, b j is the width of the radial basis function, and W i is the ith set of column vectors of the weight matrix W from the implicit layer to the output layer, which means that W = ω 11 ω 12 ω 1 m ω 21 ω 22 ω 2 m ω n 1 ω n 2 ω n m T , where i is the number of neurons in the input and output layer, and j is the number of neurons in the hidden layer.
Defining the optimal weights vector as W * , the system uncertainty part function f ( x ) can be expressed as [36]
f x = W * T H + ε
where ε = ε 1 ε 2 ε n T is the smallest approximation error of the RBF network, H = x , c , b , · is the arithmetic function from the input layer to the output of the implicit layer; since the system uncertainty f ( x ) is bounded, there exists an upper bound W m a x for W * , which means that W * W max .
Defining the estimate of the RBF network for the system uncertainty f ( x ) as f ^ x = W ^ T H , the estimation error can be written as [36]
f ˜ x = f x f ^ x = W * T H W ^ T H + ε = W * T W ^ T H + ε = W ˜ T H + ε
where ε satisfies ε ε max , with ε max a bounded constant.
Based on the fractional-order disturbance observer’s estimate of the disturbance signal d ^ and the RBF network’s estimate of the system’s uncertainty f ^ ( x ) , the sliding-mode surface of the fractional-order sliding-mode controller is given by
s = C 1 e + C 2 D λ 1 e
where e is the system tracking error, e = y d y , determined by the desired output y d and the actual output y of the system, C 1 and C 2 are positive real numbers, and 0 < λ < 1 is the order of the fractional-order differentiation.
Deriving the sliding-mode surface and taking the sliding-mode convergence law as k s k s sgn s , where k s and k are the sliding-mode convergence law parameters and k s , k > 0 , and substituting into (1), the control law can be obtained by
u = C B 1 y ˙ d + C 2 C 1 D λ e + k s + k s sgn s C 1 B 1 A x B 1 d B 1 f x
Substituting the disturbance signal estimator of (3) with the system uncertainty estimator f ^ ( x ) yields the following final control law:
u = C B 1 y ˙ d + C 2 C 1 D λ e + k s + k s sgn s C 1 B 1 A x B 1 d ^ B 1 f ^ x

3. TD3 Algorithm Based on Fractional-Order Control Method

In this section, a TD3 algorithm based on the designed fractional-order control method is proposed, and we design the reward function and loss function of the TD3 algorithm based on the proposed fractional-order control method, which in turn makes it effective for the optimization of the parameters of the fractional-order control method.
As a kind of reinforcement learning algorithm, the basic principle of the TD3 algorithm is also composed of a critic and an actor, two kinds of networks of the agent and the environment for the interactive operation. More specifically, Figure 2 shows the agent interacting with the environment to obtain the state s t and rewards r t , the output of the action a t at moment t, and the update of the two kinds of networks within the agent is based on the loss function.
In this paper, the state of the environment s t is set as the error signal for each state of the system, which means that s t = e 1 t e 2 t e n t T , and the reward signal r t is set to the negative of the weighted sum of the absolute values of the error signals for each state of the system, which means r t = i = 1 n η i e i t , where η i > 0 denotes the weight of the ith error signal in the reward function, which needs to be set according to the actual situation; the action a t of the actor network’s output are the parameters of the fractional-order controller which should be optimized.
In the reinforcement learning algorithm, the Bellman equation is defined as [37]
U t = r t + γ U t + 1
where r t denotes the reward acquired by the intelligence in the environment, U t = k = 1 n γ k 1 r t + k 1 is a value function that represents the sum of the rewards that the intelligence expects to acquire in a continuous process after moment t, and where 0 < γ < 1 denotes the discount rate.
For both sides of (9), the expectation based on the state S and action A at a moment can be obtained, and the value function of the action at that moment can be obtained by
E π U t S t = s t , A t = a t = E π r s t , a t + γ E π U t + 1 S t + 1 = s t + 1 , A t + 1 = a t + 1
In the reinforcement learning algorithm, the estimation of the value function of the action is carried out through the critic network, so there is a certain bias, at which point the Bellman equation’s equal sign does not hold. Defining the difference between the left and right sides of (10) as the temporal-difference error (TD error) δ T D t , it can be written as [38]
δ T D t = Q π s t , a t ϖ t r s t , a t + γ Q π s t + 1 , a t + 1 ϖ t
where ϖ t is the parameter of the critic network at moment t.
In order to solve the bootstrapping of the overestimation in traditional reinforcement learning algorithms, the TD3 algorithm adopts two sets of critic networks and target critic networks with identical structure and parameters; the corresponding network outputs are called Critic1 network Q 1 , π s , a ϖ 1 , Critic2 network Q 2 , π s , a ϖ 2 , Target Critic1 network Q 1 , π s , a ϖ 1 , and Target Critic2 network Q 2 , π s , a ϖ 2 . The algorithm adopts the strategy of smoothing regularization to reduce the estimation error, adding a noise signal ϵ c l i p N 0 , σ ; c , c to the action output from the actor network, so that the target value y t a r g e t = r s t , a t + γ · Q π s t + 1 , a t + 1 can be expressed as [31]
y t a r g e t = r s t , a t + γ min i = 1 , 2 Q 1 , π s t + 1 , a t + 1 + ϵ ϖ j
The agent in the TD3 algorithm stores each state transfer pair s , a , r , s in the replay buffer during each interaction with the environment, and since in the process of learning by the intelligent body, the selection of the experience should be considered to prioritize the learning value, in this paper, a nonuniform sampling was designed as
P X = r i = e α r i i = 1 n e α r i 1
where P X = r i denotes the probability of drawing a state transfer pair corresponding to reward r i , α is a parameter that controls the shape of the distribution, and n denotes the total number of samples in the replay buffer.
Based on the sampling probability in (13), N samples are taken from the replay buffer and the loss function of the critic network is constructed using the mean squared error (MSE), which can be written as
J ϖ t = j = 1 2 J ϖ j t = N 1 j = 1 2 i = 1 N Q j , π s , a ϖ i y t a r g e t j , i 2
For the actor network, in order to maximize the future expected return of the current strategy, which means that π a θ = arg max θ E s t , a t i = t γ i t R i , and in order to improve the explorability of the action space, the loss function was designed as
J θ t = N 1 j = 1 N E s t j , a t j i = t j γ i t j R i + α H θ π
where θ is the parameter of the actor network, H θ π = a π s θ log π s θ is a measure of uncertainty in the distribution of strategies, and α is the weight parameter for that item.
For both critic network and actor network parameters ϖ , θ are updated using the gradient descent method, which can be expressed as [31]
ϖ t + 1 = ϖ t α ϖ J ϖ t θ t + 1 = θ t β θ J θ t = θ t β N 1 a Q π s , a ω θ π s θ + α θ H θ π
where α and β are the learning rate of the critic network and actor network.
For the target network parameter, a soft update is performed by introducing the update shift τ of the network parameter, which means [31]:
ϖ i , t + 1 = τ ϖ i , t + 1 + 1 τ ϖ i , t i = 1 , 2 θ t + 1 = τ θ t + 1 + 1 τ θ t
Finally, the TD3 algorithm uses lagged updating for the actor network, which means that the actor network is updated once when the critic network is updated multiple times. Based on the above description, a structure of a fractional-order control method based on the TD3 algorithm is proposed, and the corresponding pseudocode is shown in Algorithm 1.
Algorithm 1 FOCS pseudocode based on the TD3 algorithm
  Initialize critic networks Q 1 , π s , a ϖ 1 and Q 2 , π s , a ϖ 2 and actor network π θ with
  random parameters ϖ 1 , ϖ 2 , θ
  Initialize target networks ϖ 1 ϖ 1 , ϖ 2 ϖ 2 , θ θ
  Initialize replay buffer B
  for t = t 0 to t = t f do
    Output action based on current parameters a t π θ s t + ϵ , ϵ c l i p N 0 , σ ; c , c ,
    Take a t for FOCS, observe the error e ( t ) , calculate r, and observe new state s t + 1
    Store the transfer pair s , a , r , s in B
    if  n N    n is the number of transfer pair in replay buffer B .
      Sample N transfer pairs and calculate the loss function of critic network J ϖ t
      Update ϖ by ϖ t + 1 ϖ t α ϖ J ϖ t
         if Critic network after d updates
           Calculate the loss function of critic network J θ t
           Update θ by θ t + 1 θ t β θ J θ t
           Update target network by
                ϖ t + 1 τ ϖ t + 1 + 1 τ ϖ t
                θ t + 1 τ θ t + 1 + 1 τ θ t
         end if
    end if
  end for

4. Stability Analysis

In this section, the stability of the proposed control method is substantiated through a detailed proof, emphasizing the theoretical feasibility of the proposed approach.
For the proposed fractional-order disturbance observer, the Lyapunov function is defined as
V d = 1 2 e d T e d
The derivation of (18) and substitution into (1), (3), and (5) yield
V ˙ d = e d T e ˙ d = e d T d ˙ d ^ ˙ = e d T d ˙ z ˙ + L x ˙ = e d T d ˙ D λ + 1 D λ z + L A x + B u + f x + d = e d T d ˙ L e d L f ˜ x = e d T M e d + e d T M f ˜ x + e d T d ˙ λ max M e d T e d + λ max M e d T W ˜ T H + λ max M e d T ε + 1 2 e d T e d + d ˙ T d ˙ 2 λ max M + 1 2 e d T e d + λ max M 2 t r H T H t r W ˜ T W ˜ + 1 2 λ max M ε T ε + d ˙ T d ˙ 2 λ max M + 1 2 e d T e d + λ max M m 2 t r W ˜ T W ˜ + 1 2 λ max M ε max 2 + ζ 2
where λ max ( M ) is the largest eigenvalue of M which is a Hurwitz matrix so it can be obtained that λ max ( M ) < 0 . In order to make the estimation error of the fractional-order disturbance observer for the disturbance signal stable, M should satisfy λ max ( M ) < 1 / 4 .
In order to keep the approximation error of the designed RBF network for the system uncertainty within a certain range, the adaptive law of the RBF network weights and the implicit layer function update law are taken as
W ^ ˙ = η 1 K W ^ C 1 H s T C
where η R m × m is the designed positive-definite diagonal matrix, K > 0 are the designed parameters of the adaptive law, s and C 1 are the sliding-mode surface and the parameter which was designed in (6), and H is the implicit layer function vector.
For the proposed RBF network, the Lyapunov function is defined as
V f = 1 2 t r ( W ˜ T η W ˜ )
Derive equation (21), and substitute into the RBF network’s adaptive law (20):
V ˙ f = t r W ˜ T η W ˜ ˙ = t r W ˜ T η W ˙ * W ^ ˙ = t r W ˜ T η η 1 K W ^ C 1 H s T C = t r W ˜ T K W ^ + t r W ˜ T C 1 H s T C = K t r W ˜ T W * W ˜ + t r W ˜ T C 1 H s T C K 2 t r W ˜ T W ˜ + K 2 W * 2 + t r W ˜ T C 1 H s T C K 2 t r W ˜ T W ˜ + K 2 W max 2 + t r W ˜ T C 1 H s T C
Based on the designed sliding-mode surface function, define the Lyapunov function
V s = 1 2 s T s
Deriving equation (23) and substituting the system state equation shown in (1) with the control law shown in (8) yield
V ˙ s = s T s ˙ = s T C 1 e ˙ + C 2 D λ e = s T C 1 y ˙ d C x ˙ + C 2 D λ e = s T C 1 y ˙ d C A x + B u + f x + d + C 2 D λ e = s T C 1 y ˙ d C A x + B C B 1 y ˙ d + k s + k s sgn s + C 2 D λ e C 1 B 1 A x B 1 d ^ B 1 f ^ x + f x + d + C 2 D λ e = s T k s k s sgn s C 1 C f ˜ x C 1 C e d = k s T s s T k s sgn s + C 1 C f ˜ x + C 1 C e d k + k s 2 + C 1 C 2 s T s + C 1 C 2 e d T e d + k s 2 s T C 1 C W ˜ T H + ε
To prove the overall stability of the original system with the addition of a fractional-order control method, define the Lyapunov function
V = V d + V f + V s
A derivation of equation (25) and a substitution of the Lyapunov derivation of (19), (22), and (24) lead to
V ˙ = V d . + V f . + V s . 2 λ max M + C 1 C + 1 2 e d T e d + k + k s 2 + C 1 C 2 s T s K 2 m λ max M 2 t r W ˜ T W ˜ + C 1 C 2 s T s + C 1 C 2 ε max 2 + K 2 W max 2 + 1 2 ζ 2 + λ max M ε max 2 + k s 2 k + k s 2 + C 1 C s T s + 2 λ max M + C 1 C + 1 2 e d T e d K 2 m λ max M 2 t r W ˜ T W ˜ + 1 2 ζ 2 + λ max M ε max 2 + C 1 C ε max 2 + K W max 2 + k s k k s 2 C 1 C s T s 2 λ max M C 1 C + 1 2 e d T e d K 2 m λ max M 2 t r W ˜ T W ˜ + δ κ V + δ
where κ = min k k s 2 C 1 C , 2 λ max M C 1 C + 1 2 , K 2 m λ max M 2 to satisfy the system stability, the control system parameter 0 < C 1 < 4 λ max M 1 C , k k s 2 > C 1 C , k s > 0 should be selected, and δ = 1 2 ζ 2 + λ max M ε max 2 + C 1 C ε max 2 + K W max 2 + k s . At this point, the system is boundedly stable.

5. Simulation Result

In this section, in order to validate the effectiveness of the proposed methodology, a valve-controlled hydraulic system is selected as the object of study to design a series of simulations. Firstly, to verify the online learnability of the TD3 algorithm, three rounds of training states during the training process are selected for the comparison. Secondly, the control effect is verified for the optimized fractional-order control method of the TD3 algorithm by employing the prescribed performance fractional-order sliding-mode controller (PPC-FOSMC), an unoptimized fractional-order sliding-mode controller (FOSMC), and a sliding-mode controller (SMC). Furthermore, the antidisturbance is proved for the proposed method, and the simulation verification is carried out for different input signals and disturbance signals. The results show that the fractional-order control method based on the TD3 algorithm not only has good online learning ability and generalization ability but also has a better control effect than the traditional FOSMC, SMC, and PPC-FOSM. Finally, in order to verify that the designed method can still maintain a better control effect under noise disturbance, a Gaussian noise is selected as the disturbance signal to simulate using the online learning method.

5.1. Simulation Results of System Online Learning

A valve-controlled hydraulic system is a third-order system [39] which can be described approximately as
x ˙ = A x + B u + B d d + f x y = C x
where x = x v , x s , x a T denotes the position, velocity, and acceleration, respectively, u denotes the control input, and y denotes the output of the position. The state matrix, input matrix, disturbance matrix, and output matrix A , B , B d , and C are, respectively,
A = 0 1 0 0 0 1 R 1 R 2 R 3 , B = 0 0 b , B d = 0 0 1 T , C = 1 0 0 ,
where R 1 = 4 k k c e β m V t , R 2 = 4 B p k c e β m V t + k m , R 3 = 4 k c e β V t + B p m , b = 4 k v k s v k q β A p m V t , and the values of the parameters of the valve-controlled hydraulic system are shown in Table 1.
The disturbance signal during the TD3 training was set to be d = 0.5 s i n ( 5 π t ) , the system uncertainty was f ( x ) = s i n ( 2 x a ) c o s ( 3 x a ) , and the system expected the output position signal to be x p d = s i n ( t ) ; the training parameters are shown in Table 2.
Learning to train the system with the above parameters, the episode rewards and average rewards obtained after 1900 episodes are shown in Figure 3a, and Figure 3b–d show the state of the system for the selected episode 1, episode 500, and episode 1826. It can be noticed that in the first 420 episodes, the episode reward fluctuates a lot, and after episode 420, the episode reward and the average reward stabilize for the first time, but at that point, the network still cannot achieve a particularly good training effect. From episode 630 to episode 1300, the agent further explores the action space, and after episode 1300, the episode reward and average reward converge to more optimal values, corresponding to the system state. It can be found that the position and velocity signals of the system corresponding to episode 1 have the worst tracking performance with both large errors and desired signals, while the acceleration signals have the worst tracking performance with more noise. Compared to episode 1, episode 500’s systematic position signal can already track the desired signal at the 0.8th second but with a small error, the velocity signal can track the desired signal after 2 s but with a small noise in between, and the acceleration signal can only track the desired signal after the 4th second. Finally, compared to the previous two, episode 1826 is definitely the best performer, with the position signal tracking the desired signal in the first 0.1 s with very little error, and the velocity signal tracking the desired signal with little error after that, despite oscillating in the first 0.1 s, and with the acceleration signal tracking the desired signal after oscillating in the first 0.4 s with an acceptable error.
In order to better visually represent the training state of set 1826, the system’s individual signals and desired signals and their errors at that point in time are shown in Figure 4. It can be seen that the position, velocity, and acceleration signals of the system have good tracking performance, and the errors are all within a small range. The same is true for the estimation of disturbance signals.

5.2. Simulation Results of TD3-FOCS

For the online learning results of the valve-controlled hydraulic fractional-order control method, the training results of the 1826th episode are shown in Figure 5 with C 1 = 0.001 , C 2 = 290 , k = 945 , k s = 43 , λ = 0.55 selected as the controller parameters, and the parameters of FODOB and the RBF network were set to L = 157 , K = 7 , and η = I , which is a single-unit diagonal matrix for the subsequent comparative simulation experiments; the simulation time was set to 100 s, and the final results are shown in Figure 6.
As can be seen from Figure 6, for the position signal, the proposed TD3-FOSMC had the fastest convergence speed (first 0.03 s) and maintained the final steady-state error in the range of (−0.001, +0.001), which was the best among the experimental results. For the velocity signal, it also had the fastest convergence speed (first 0.25 s) and maintained the final steady-state error in the range of ( 0.005 , + 0.005 ) , which was the best performance among all the experimental methods. For acceleration signals, although the prescribed performance of the fractional-order sliding-mode controller ended up with a smaller error range, the convergence time was greater than the proposed TD3-FOSMC, which converged in the first 0.3 s. Combining all the above control effects, it can be concluded that the proposed TD3-FOSMC has a better speed and stability.

5.3. Simulation Results of Different Control Signals and Disturbance Signals

In order to demonstrate the robustness and versatility of the method, two control signals (a triangular wave composite signal and a step signal) and two disturbance signals (a strong sine wave signal disturbance and a triangular wave composite signal) were selected for the simulation verification. This allowed a more comprehensive assessment of the method’s antidisturbance performance under different control and disturbance conditions, thus demonstrating the effectiveness of the method in mitigating the effects of an external disturbance in a variety of practical situations.
As observed in Figure 7, the simulation results revealed the capability of accurately tracking the desired signal under the influence of a triangular wave composite signal and a strong sine wave signal disturbance. Specifically, the position signal achieved tracking within a remarkably short duration of 0.07 s, with very small overshoots (less than 1%) in the first 0.02 s, and the final steady-state error confined to a narrow range of (−0.001, 0.001). In the first 0.3 s, the velocity signal had large to small fluctuations in the range of (−3.7, 6); after that, it achieved synchronization with the desired signal and kept the steady-state error within the range of (−0.01, 0.01). On the other hand, the acceleration signal fluctuated in the range of (−12, 12) for the first 0.5 s and was after synchronized with the desired signal and kept the steady-state error within the range of (−0.25, 0.25). The results show that despite the large overshoot in the system response for the different input signals and disturbance signals, the system still had good stability and fast performance.
Analyzing the results shown in Figure 8, when subjected to a step signal and a strong triangular wave composite signal disturbance during the simulation, the obtained results exhibited a marginally reduced performance compared to the previous scenario. Taking the simulation time of 10 s as an example, it can be found that for the position, velocity, and acceleration signals, there were large fluctuations in the first 1 s, and their stabilization time was slowing down sequentially. For the position signal, there were fluctuations within the range of (5, 15) within the first 0.3 s; after that, it could track the desired signal, and the final steady-state error was limited to the range of (0, 0.004). The velocity signal exhibited fluctuations within the range of (−10, 10) during the first 0.5 s, after which it synchronized with the desired signal, and the final steady-state error was constrained to (−0.002, 0). For the acceleration signal, fluctuations within the range of (−20, 20) occurred during the initial 0.8 s, following which it converged to the desired value. The final steady-state error remained confined to a smaller range of (−0.0001, 0.0001). Despite a slight decrease in performance, the results underscore the method’s ability to maintain effective control even in the presence of challenging input signals and disturbance conditions.

5.4. Simulation Results under Noise Disturbance

The control performance is illustrated for the method under noise disturbance. The noise disturbance signal was selected with a mean value of zero, a variance of five, and a frequency of 10 Hz, and the simulation verification was carried out by using the online learning of the agent. The results are shown in Figure 9.
Analyzing Figure 9, it can be seen that the position signal exhibited minimal susceptibility form the noise disturbance signal, with an overshoot of less than 3% observed only within the initial 0.03 s. Subsequently, the position signal could track the desired signal, and the final steady-state error could be maintained within the range of (−0.001, 0). The velocity signal demonstrated minimal susceptibility to noise disturbances, with oscillations confined to the range of (−1.6, 4.7) within the initial 0.08 s. Following that period, the signal could adeptly track the desired trajectory, and the final steady-state error could satisfy the narrow bounds of (−0.0015, 0.0015). Comparing with the earlier online learning outcomes, the performance of the acceleration signal was degraded under the noise signal. The acceleration signal was oscillatory within the range of (−6, 6) during the initial 0.15 s. Despite these challenges, it could approximately track the desired signal. However, due to the influence of noise, it maintained a final steady-state error within the range of (−0.13, 0.13). Thus, based on the analysis above, the proposed control method could achieve the control of system within the bounded range under the noise signal.

6. Conclusions

In this paper, a fractional-order control method based on the TD3 algorithm was introduced. A fractional-order disturbance observer was designed to estimate the system’s disturbance signal, and an RBF network was selected to approximate the uncertainties in the system, and the fractional-order sliding-mode control method was also adopted to design the controller. A valve-controlled hydraulic system was simulated and validated in Matlab/Simulink using the agent online learning and the optimization parameters. Different control signals and disturbance signals were used in the optimized fractional-order control system. The results showed that the limitations of this method mainly lay in the difficulty of setting up the training environment comprehensively. Despite some shortcomings, the proposed method was generally fast and had good antidisturbance ability.

Author Contributions

Conceptualization, G.J. and S.S.; methodology, G.J. and Z.A.; software, G.J. and D.S.; validation, G.J. and D.S.; formal analysis, G.J. and Z.A.; writing—original draft preparation, G.J.; writing—review and editing, G.J. and S.S.; supervision, S.S.; funding acquisition, S.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Fund of Science and Technology on Space Intelligent Control Laboratory Foundation (grant number HTKJ2023KL502002); the China Postdoctoral Science Foundation (grant number 2020M681587); the Jiangsu Province Postdoctoral Science Foundation (grant number 2020Z112).

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Spanos, P.D.; Malara, G. Nonlinear vibrations of beams and plates with fractional derivative elements subject to combined harmonic and random excitations. Probabilistic Eng. Mech. 2020, 59, 142–149. [Google Scholar] [CrossRef]
  2. Magin, R.L. Fractional calculus models of complex dynamics in biological tissues. Comput. Math. Appl. 2010, 36, 1586–1593. [Google Scholar] [CrossRef]
  3. Atangana, A. Non validity of index law in fractional calculus: A fractional differential operator with Markovian and non-Markovian properties. Phys. A Stat. Mech. Its Appl. 2018, 505, 688–706. [Google Scholar] [CrossRef]
  4. Gonzalez, E.A.; Petráš, I. Advances in fractional calculus: Control and signal processing applications. In Proceedings of the 2015 16th International Carpathian Control Conference (ICCC), Szilvasvarad, Hungary, 27–30 May 2015; pp. 147–152. [Google Scholar] [CrossRef]
  5. Warrier, P.; Shah, P. Optimal Fractional PID Controller for Buck Converter Using Cohort Intelligent Algorithm. Appl. Syst. Innov. 2021, 4, 50. [Google Scholar] [CrossRef]
  6. Razzaghian, A. A fuzzy neural network-based fractional-order Lyapunov-based robust control strategy for exoskeleton robots: Application in upper-limb rehabilitation. Math. Comput. Simul. 2022, 193, 567–583. [Google Scholar] [CrossRef]
  7. Fei, J.; Wang, Z.; Pan, Q. Self-Constructing Fuzzy Neural Fractional-Order Sliding Mode Control of Active Power Filter. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 10600–10611. [Google Scholar] [CrossRef]
  8. Jakovljević, B.; Pisano, A.; Rapaić, M.R.; Usai, E. On the sliding-mode control of fractional-order nonlinear uncertain dynamics. Int. J. Robust Nonlinear Control 2016, 24, 782–798. [Google Scholar] [CrossRef]
  9. Mirrezapour, S.Z.; Zare, A.; Hallaji, M. A new fractional sliding mode controller based on nonlinear fractional-order proportional integral derivative controller structure to synchronize fractional-order chaotic systems with uncertainty and disturbances. J. Vib. Control 2022, 28, 773–785. [Google Scholar] [CrossRef]
  10. Deepika, D. Hyperbolic uncertainty estimator based fractional-order sliding mode control framework for uncertain fractional-order chaos stabilization and synchronization. ISA Trans. 2022, 123, 76–86. [Google Scholar] [CrossRef] [PubMed]
  11. Delavari, H.; Ghaderi, R.; Ranjbar, A.; Momani, S. Fuzzy fractional-order sliding mode controller for nonlinear systems. Commun. Nonlinear Sci. Numer. Simul. 2010, 15, 963–978. [Google Scholar] [CrossRef]
  12. Sun, G.; Ma, Z. Practical tracking control of linear motor with adaptive fractional-order terminal sliding mode control. IEEE/ASME Trans. Mechatron. 2017, 22, 2643–2653. [Google Scholar] [CrossRef]
  13. Djari, A.; Bouden, T.; Boulkroune, A. Design of fractional-order sliding mode controller (FSMC) for a class of fractional-order non-linear commensurate systems using a particle swarm optimization (PSO) Algorithm. J. Control Eng. Appl. Inform. 2014, 16, 46–55. [Google Scholar]
  14. Karthikeyan, A.; Rajagopal, K. Chaos control in fractional-order smart grid with adaptive sliding mode control and genetically optimized PID control and its FPGA implementation. Complexity 2017, 2017, 3815146. [Google Scholar] [CrossRef]
  15. Han, S. Modified grey-wolf algorithm optimized fractional-order sliding mode control for unknown manipulators with a fractional-order disturbance observer. IEEE Access 2020, 8, 18337–18349. [Google Scholar] [CrossRef]
  16. Salman, G.A.; Jafar, A.S.; Ismael, A.I. Application of artificial intelligence techniques for LFC and AVR systems using PID controller. Int. J. Power Electron. Drive Syst. 2019, 10, 1694. [Google Scholar] [CrossRef]
  17. Perrusquía, A.; Yu, W. Identification and optimal control of nonlinear systems using recurrent neural networks and reinforcement learning: An overview. Neurocomputing 2021, 438, 145–154. [Google Scholar] [CrossRef]
  18. Liu, Y.J.; Zhao, W.; Liu, L.; Li, D.; Tong, S.; Chen, C.P. Adaptive neural network control for a class of nonlinear systems with function constraints on states. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 2732–2741. [Google Scholar] [CrossRef] [PubMed]
  19. Chen, Z.; Niu, B.; Zhang, L.; Zhao, J.; Ahmad, A.M.; Alassafi, M.O. Command filtering-based adaptive neural network control for uncertain switched nonlinear systems using event-triggered communication. Int. J. Robust Nonlinear Control 2022, 32, 6507–6522. [Google Scholar] [CrossRef]
  20. Katz, S.M.; Corso, A.L.; Strong, C.A.; Kochenderfer, M.J. Verification of image-based neural network controllers using generative models. J. Aerosp. Inf. Syst. 2022, 19, 574–584. [Google Scholar] [CrossRef]
  21. Kiumarsi, B.; Vamvoudakis, K.G.; Modares, H.; Lewis, F.L. Optimal and autonomous control using reinforcement learning: A survey. IEEE Trans. Neural Netw. Learn. Syst. 2017, 29, 2042–2062. [Google Scholar] [CrossRef]
  22. Duraisamy, P.; Nagarajan Santhanakrishnan, M.; Rengarajan, A. Design of deep reinforcement learning controller through data-assisted model for robotic fish speed tracking. J. Bionic Eng. 2023, 20, 953–966. [Google Scholar] [CrossRef]
  23. Wei, C.; Xiong, Y.; Chen, Q.; Xu, D. On adaptive attitude tracking control of spacecraft: A reinforcement learning based gain tuning way with guaranteed performance. Adv. Space Res. 2023, 71, 4534–4548. [Google Scholar] [CrossRef]
  24. Barto, A.G. Reinforcement learning. In Neural Systems for Control; Academic Press: Cambridge, MA, USA, 1997; pp. 7–30. [Google Scholar]
  25. Nian, R.; Liu, J.; Huang, B. A review on reinforcement learning: Introduction and applications in industrial process control. Comput. Chem. Eng. 2020, 139, 106886. [Google Scholar] [CrossRef]
  26. Zhu, Y.; Zhao, D.; Li, X. Using reinforcement learning techniques to solve continuous-time non-linear optimal tracking problem without system dynamics. IET Control Theory Appl. 2016, 10, 1339–1347. [Google Scholar] [CrossRef]
  27. Syafiie, S.; Tadeo, F.; Martinez, E. Model-free learning control of neutralization processes using reinforcement learning. Eng. Appl. Artif. Intell. 2007, 20, 767–782. [Google Scholar] [CrossRef]
  28. Dogru, O.; Velswamy, K.; Ibrahim, F.; Wu, Y.; Sundaramoorthy, A.S.; Huang, B.; Xu, S.; Nixon, M.; Bell, N. Reinforcement learning approach to autonomous PID tuning. Comput. Chem. Eng. 2022, 161, 107760. [Google Scholar] [CrossRef]
  29. Taherian, S.; Kuutti, S.; Visca, M.; Fallah, S. Self-adaptive torque vectoring controller using reinforcement learning. In Proceedings of the 2021 IEEE International Intelligent Transportation Systems Conference (ITSC), Indianapolis, IN, USA, 19–22 September 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 172–179. [Google Scholar]
  30. Yan, L.; Webber, J.L.; Mehbodniya, A.; Moorthy, B.; Sivamani, S.; Nazir, S.; Shabaz, M. Distributed optimization of heterogeneous UAV cluster PID controller based on machine learning. Comput. Electr. Eng. 2022, 101, 108059. [Google Scholar] [CrossRef]
  31. Wu, J.; Wu, Q.J.; Chen, S.; Pourpanah, F.; Huang, D. A-TD3: An Adaptive Asynchronous Twin Delayed Deep Deterministic for Continuous Action Spaces. IEEE Access 2022, 10, 128077–128089. [Google Scholar] [CrossRef]
  32. Dankwa, S.; Zheng, W. Twin-delayed ddpg: A deep reinforcement learning technique to model a continuous movement of an intelligent robot agent. In Proceedings of the 3rd International Conference on Vision, Image and Signal Processing, Vancouver, BC, Canada, 26–28 August 2019; pp. 1–5. [Google Scholar]
  33. Li, C.; Qian, D.; Chen, Y.Q. On Riemann-Liouville and caputo derivatives. Discret. Dyn. Nat. Soc. 2011, 2011, 562494. [Google Scholar] [CrossRef]
  34. Walsh, G.C.; Ye, H.; Bushnell, L.G. Stability analysis of networked control systems. IEEETrans. Control Syst. Technol. 2002, 10, 438–446. [Google Scholar] [CrossRef]
  35. Er, M.J.; Wu, S.; Lu, J.; Toh, H.L. Face recognition with radial basis function (RBF) neural networks. IEEE Trans. Neural Netw. 2002, 13, 697–710. [Google Scholar] [PubMed]
  36. Fei, J.; Wang, H.; Fang, Y. Novel neural network fractional-order sliding-mode control with application to active power filter. IEEE Trans. Syst. Man, Cybern. Syst. 2021, 52, 3508–3518. [Google Scholar] [CrossRef]
  37. Fei, Y.; Yang, Z.; Chen, Y.; Wang, Z. Exponential bellman equation and improved regret bounds for risk-sensitive reinforcement learning. Adv. Neural Inf. Process. Syst. 2021, 34, 20436–20446. [Google Scholar]
  38. Doya, K. Reinforcement learning in continuous time and space. Neural Comput. 2000, 12, 219–245. [Google Scholar] [CrossRef]
  39. Duan, Z.; Sun, C.; Li, J.; Tan, Y. Research on servo valve-controlled hydraulic motor system based on active disturbance rejection control. Meas. Control. 2024, 57, 113–123. [Google Scholar] [CrossRef]
Figure 1. Structure of the RBF network.
Figure 1. Structure of the RBF network.
Fractalfract 08 00099 g001
Figure 2. The relationship between environment and agent.
Figure 2. The relationship between environment and agent.
Fractalfract 08 00099 g002
Figure 3. Episode reward and average reward with the training results of the 1826th episode: (a) episode reward and average reward of 1900 episodes, (b) the 1st, 500th, and 1826th episode’s position and expected position signals; (c) the 1st, 500th, and 1826th episode’s speed and expected signals; (d) the 1st, 500th, 1826th episode’s acceleration and expected signals.
Figure 3. Episode reward and average reward with the training results of the 1826th episode: (a) episode reward and average reward of 1900 episodes, (b) the 1st, 500th, and 1826th episode’s position and expected position signals; (c) the 1st, 500th, and 1826th episode’s speed and expected signals; (d) the 1st, 500th, 1826th episode’s acceleration and expected signals.
Fractalfract 08 00099 g003
Figure 4. The 1st, 500th, and 1826th episode’s system signal and expected signals with the error: (a) position signal and expected signal, (b) position tracking error, (c) speed signal and expected signal, (d) speed tracking error, (e) acceleration signal and expected signal, (f) acceleration tracking error, (g) estimation of the disturbance signal and the actual signal, (h) estimation error of the disturbance signal.
Figure 4. The 1st, 500th, and 1826th episode’s system signal and expected signals with the error: (a) position signal and expected signal, (b) position tracking error, (c) speed signal and expected signal, (d) speed tracking error, (e) acceleration signal and expected signal, (f) acceleration tracking error, (g) estimation of the disturbance signal and the actual signal, (h) estimation error of the disturbance signal.
Fractalfract 08 00099 g004
Figure 5. Training results of the 1826th episode: (a) parameter C 1 , (b) parameter C 2 , (c) parameter k, (d) parameter k s .
Figure 5. Training results of the 1826th episode: (a) parameter C 1 , (b) parameter C 2 , (c) parameter k, (d) parameter k s .
Fractalfract 08 00099 g005
Figure 6. TD3-optimized fractional-order control vs. unoptimized control method’s system signals and expected signals and their errors: (a) position signal and expected signal, (b) position tracking error, (c) speed signal and expected signal, (d) speed tracking error, (e) acceleration signal and expected signal, (f) acceleration tracking error.
Figure 6. TD3-optimized fractional-order control vs. unoptimized control method’s system signals and expected signals and their errors: (a) position signal and expected signal, (b) position tracking error, (c) speed signal and expected signal, (d) speed tracking error, (e) acceleration signal and expected signal, (f) acceleration tracking error.
Fractalfract 08 00099 g006
Figure 7. Simulation results of triangular wave composite signals and strongly interfering sine wave signals: (a) position signal and expected signal, (b) position tracking error, (c) speed signal and expected signal, (d) speed tracking error, (e) acceleration signal and expected signal, (f) acceleration tracking error, (g) estimation of the disturbance signal and actual signal, (h) estimation error of the disturbance signal.
Figure 7. Simulation results of triangular wave composite signals and strongly interfering sine wave signals: (a) position signal and expected signal, (b) position tracking error, (c) speed signal and expected signal, (d) speed tracking error, (e) acceleration signal and expected signal, (f) acceleration tracking error, (g) estimation of the disturbance signal and actual signal, (h) estimation error of the disturbance signal.
Fractalfract 08 00099 g007
Figure 8. Simulation results of step signals and triangular wave composite signal: (a) position signal and expected signal, (b) position tracking error, (c) speed signal and expected signal, (d) speed tracking error, (e) acceleration signal and expected signal, (f) acceleration tracking error, (g) estimation of the disturbance signal and actual signal, (h) estimation error of the disturbance signal.
Figure 8. Simulation results of step signals and triangular wave composite signal: (a) position signal and expected signal, (b) position tracking error, (c) speed signal and expected signal, (d) speed tracking error, (e) acceleration signal and expected signal, (f) acceleration tracking error, (g) estimation of the disturbance signal and actual signal, (h) estimation error of the disturbance signal.
Fractalfract 08 00099 g008
Figure 9. Simulation results under noise disturbance: (a) position signal and expected signal, (b) position tracking error, (c) speed signal and expected signal, (d) speed tracking error, (e) acceleration signal and expected signal, (f) acceleration tracking error.
Figure 9. Simulation results under noise disturbance: (a) position signal and expected signal, (b) position tracking error, (c) speed signal and expected signal, (d) speed tracking error, (e) acceleration signal and expected signal, (f) acceleration tracking error.
Fractalfract 08 00099 g009
Table 1. Parameters of the valve-controlled hydraulic system.
Table 1. Parameters of the valve-controlled hydraulic system.
NameSymbolNumerical Value
Total mass of piston and load converted to pistonm5 kg
Piston area of hydraulic cylinder A p 0.4734 m2
Springinessk1
Flow-pressure amplification factor k c e 3.005 × 10 13
Modulus of elasticity of hydraulic fluid β 0.005
Total volume of hydraulic cylinder V t 1.5 m3
Viscous damping coefficient of piston and load B p 5
Servo amplifier gain k v 1
Servo valve gain k s v 4.733 × 10 3
Flow gain of slide valve k q 1
Table 2. Parameters of the TD3 algorithm.
Table 2. Parameters of the TD3 algorithm.
NameNumerical Value
Episode1300
Episode step100
Sampling time 0.1 s
Final time10 s
Learning rate0.001
Replay buffer 1 × 10 6
Minimum batch size128
Number of actor network layers4
Number of critic network layers9
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jiao, G.; An, Z.; Shao, S.; Sun, D. Fractional-Order Control Method Based on Twin-Delayed Deep Deterministic Policy Gradient Algorithm. Fractal Fract. 2024, 8, 99. https://doi.org/10.3390/fractalfract8020099

AMA Style

Jiao G, An Z, Shao S, Sun D. Fractional-Order Control Method Based on Twin-Delayed Deep Deterministic Policy Gradient Algorithm. Fractal and Fractional. 2024; 8(2):99. https://doi.org/10.3390/fractalfract8020099

Chicago/Turabian Style

Jiao, Guangxin, Zhengcai An, Shuyi Shao, and Dong Sun. 2024. "Fractional-Order Control Method Based on Twin-Delayed Deep Deterministic Policy Gradient Algorithm" Fractal and Fractional 8, no. 2: 99. https://doi.org/10.3390/fractalfract8020099

Article Metrics

Back to TopTop