Next Article in Journal
Effects of Colloid Milling and Hot-Water Pretreatment on Physical Properties and Enzymatic Digestibility of Oak Wood
Next Article in Special Issue
An End-to-End Deep Learning Method for Voltage Sag Classification
Previous Article in Journal
Fast Experimental Magnetic Model Identification for Synchronous Reluctance Motor Drives
Previous Article in Special Issue
Analysis and Design of a High-Efficiency SiC MOSFET 6-Phase Boost Rectifier
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improvement of PMSM Sensorless Control Based on Synergetic and Sliding Mode Controllers Using a Reinforcement Learning Deep Deterministic Policy Gradient Agent †

by
Marcel Nicola
1,
Claudiu-Ionel Nicola
1,2,* and
Dan Selișteanu
2
1
Research and Development Department, National Institute for Research, Development and Testing in Electrical Engineering—ICMET Craiova, 200746 Craiova, Romania
2
Department of Automatic Control and Electronics, University of Craiova, 200585 Craiova, Romania
*
Author to whom correspondence should be addressed.
The present work is an extension of the paper: “Improvement of PMSM Control Using Reinforcement Learning Deep Deterministic Policy Gradient Agent”. Presented at the Ee 2021, Novi Sad, Serbia, 27–30 October 2021.
Energies 2022, 15(6), 2208; https://doi.org/10.3390/en15062208
Submission received: 27 January 2022 / Revised: 28 February 2022 / Accepted: 16 March 2022 / Published: 17 March 2022
(This article belongs to the Special Issue Recent Advances in Smart Power Electronics)

Abstract

:
The field-oriented control (FOC) strategy of a permanent magnet synchronous motor (PMSM) in a simplified form is based on PI-type controllers. In addition to their low complexity (an advantage for real-time implementation), these controllers also provide limited performance due to the nonlinear character of the description equations of the PMSM model under the usual conditions of a relatively wide variation in the load torque and the high dynamics of the PMSM speed reference. Moreover, a number of significant improvements in the performance of PMSM control systems, also based on the FOC control strategy, are obtained if the controller of the speed control loop uses sliding mode control (SMC), and if the controllers for the inner control loops of id and iq currents are of the synergetic type. Furthermore, using such a control structure, very good performance of the PMSM control system is also obtained under conditions of parametric uncertainties and significant variations in the combined rotor-load moment of inertia and the load resistance. To improve the performance of the PMSM control system without using controllers having a more complicated mathematical description, the advantages provided by reinforcement learning (RL) for process control can also be used. This technique does not require the exact knowledge of the mathematical model of the controlled system or the type of uncertainties. The improvement in the performance of the PMSM control system based on the FOC-type strategy, both when using simple PI-type controllers or in the case of complex SMC or synergetic-type controllers, is achieved using the RL based on the Deep Deterministic Policy Gradient (DDPG). This improvement is obtained by using the correction signals provided by a trained reinforcement learning agent, which is added to the control signals ud, uq, and iqref. A speed observer is also implemented for estimating the PMSM rotor speed. The PMSM control structures are presented using the FOC-type strategy, both in the case of simple PI-type controllers and complex SMC or synergetic-type controllers, and numerical simulations performed in the MATLAB/Simulink environment show the improvements in the performance of the PMSM control system, even under conditions of parametric uncertainties, by using the RL-DDPG.

1. Introduction

Interest in the study of the PMSM and its applications has increased in the past decade due to its constructive advantages, such as small size, low harmonic distortion, high torque density, and easy cooling methods. The increased interest in the PMSM is also supported by the fact that it can be used in the aerospace industry, computer peripherals, numerically controlled machines, robotics, electric drives, etc. [1,2,3,4,5,6,7].
Naturally, the study of PMSM control has also attracted, in particular, the attention of researchers involved in the study and development of applications in the above-mentioned fields. Thus, we can mention the FOC-type control structure, where, in the simplest case, a PI-type controller is used for the outer PMSM rotor speed control loop, and hysteresis ON/OFF-type controllers are used for the inner current control loop. Moreover, a number of types of control systems have been implemented for the PMSM control, including: the adaptive control [8,9,10], the predictive control [11,12,13], the robust control [14,15], the backstepping control [16], the SMC [17,18], the synergetic control [19,20], the fuzzy control [21], the neuro-fuzzy control [22], the artificial neural network control (ANN) [23,24,25], and the control based on particle swarm optimization (PSO) or genetic algorithms [26].
In addition, in terms of the PMSM sensorless control, a series of PMSM rotor speed observers have been developed in order to eliminate hardware systems specific to measuring transducers, thus providing increased system reliability. The types of observers used include Luenberger [27], model reference adaptive system (MRAS) [28], and sliding mode observer (SMO) speed observers [29,30]. Kalman-type observers are used for the systems described by stochastic means [31].
The intelligent control system has a special role in the development of and improvement in control systems. These include RL agents, which are characterized by the fact that they do not use the mathematical description of the controlled system, but use input signals that contain information on the state of the system and provide control actions to optimize a Reward that consists of signals containing information on the controlled process [32,33,34,35,36,37]. Moreover, building on four papers by the authors based on the RL agent, which demonstrates the capacity of this type of algorithm to improve the performance of the linear control algorithms of the PMSM (the PI type algorithm [26,38], LQR control algorithm [39], and Feedback Linearization control algorithm [40]), this article demonstrates that a cascade control structure containing nonlinear SMC and synergetic control algorithms can also provide improved performance by using an RL agent.
This paper is a follow-up to a paper [38] presented at the 21st International Symposium on Power Electronics (Ee 2021), which shows an improvement in the performance of the FOC-type control structure of the PMSM by using an RL agent. The type of RL used is the Twin-Delayed Deep Deterministic Policy Gradient (TD3) agent, which is an extension and improved variant of the DDPG agent [41]. Moreover, based on the fact that the FOC-type control structure, in which the controller of the outer speed control loop is of the SMC type and the controllers of the inner current control loop are of the synergetic type, provides peak performance of the control system [42,43], this paper also presents the improvement in the performance of this control structure by using an RL agent. The commonly used machine learning (ML) techniques are the following: Linear Regression, Decision Tree, Support Vector Machine, Neural Networks and Deep Learning, and Ensemble of Tree. Generally, the three main types of machine learning strategies are the following: unsupervised learning, supervised learning, and RL. First, unsupervised learning is typically utilized in the fields of data clustering and dimensionality reduction. Supervised learning deals mainly with classification and regression problems. RL is a framework for learning the relationship between states and actions. Ultimately, the agent can maximize the expected Reward and learn how to complete a mission. This is a very strong analogy with the control of an industrial process. Although the other types of ML can be used to estimate certain parameters, RL is recommended for the control of an industrial process. It is clear that RL is an especially powerful form of artificial intelligence, and we are sure to see more progress from these teams. Moreover, the RL-TD3 used in this paper is an improved version of the Deep Deterministic Policy Gradient (DDPG) agent and is considered the most suitable RL agent for industrial process control [41]. In addition, the authors’ papers [26,38,39,40] compare the results of control systems improved by using the RL-TD3 agent or optimized control law parameters by means of computational intelligence algorithms: particle swarm optimization (PSO) algorithm, simulated annealing (SA) algorithm, genetic algorithm (GA), and gray wolf optimization (GWO) algorithm. The conclusion clearly shows the superiority of the performance of the PMSM control system the in case of using the RL-TD3 agent. However, we consider that the control structures and the comparative results presented are relevant for this paper.
  • The main contributions of this paper are as follows:
  • The proposal of a PMSM control structure, where the controller of the outer rotor speed control loop is of the SMC type and the controllers for the inner control loops of id and iq currents are of the synergetic type.
  • The improvement in the performance of the PMSM control system based on the FOC-type strategy, both when using simple PI-type controllers or in the case of complex SMC or synergetic-type controllers, is achieved using the RL based on the TD3 agent.
The validations of the results are performed through numerical simulations in the MATLAB/Simulink environment to show the improvements in the performance of the PMSM control system even under the conditions of parametric uncertainties, by using the RL-TD3 agent.
The rest of the paper is organized as follows: Section 2 describes the RL for process control, Section 3 presents a correction of the control signals for the PMSM based on the FOC strategy using the RL-TD3 agent, and Section 4 presents a correction of the control signals for the PMSM FOC strategy based on SMC and synergetic control using the RL-TD3 agent. The results of the numerical simulations are presented in Section 5, and Section 6 presents conclusions.

2. Reinforcement Learning for Process Control

Generally speaking, RL consists of learning the execution of a task by the host computer if it interacts dynamically with an unknown process. Without explicit programming of the task learning method, the learning process is based on a set of decisions that are made so as to achieve the maximum of the cumulative Reward, whose expression is set a priori. Figure 1 shows the schematic diagram for a RL scenario.
The input signals to the agent are Observation and Reward. The Observation consists of a set of predefined signals characterizing the process, and the Reward is defined as a measure of the success based on the Action output signal. The Action is given by the control quantities of the controlled process. The Observations represents the signals visible for the agent and are found in the form of measured signals, their rate of change, and the related errors.
Usually, the Reward is created as part of the continuous actions in the form of a sum, such as the square of the error of the signals of interest and the square of past actions. These terms are given a weight determined by the final purpose of the problem. Usually, in the process control, the Reward is given by a function that implements the minimization of the steady-state error.
The agent contains a component called Policy, and also implements a learning algorithm. The Policy represents the way in which the actions are associated with the Observations of the process and is described by a function with adjustable parameters. In the case of the process control, the Policy is similar to the operating mode of a controller.
The learning algorithm is designed to find an optimal Policy by continuously updating the parameters of the function associated to the Policy based on the maximization of the cumulative Reward. The process (in the general case referred to as the environment) consists of a plant, reference signals and the associated errors, filters, disturbances, measurement noise, and analog-to-digital and digital-to-analog converters.
The main stages of the RL process are [35,36,37,38,41]:
  • Problem statement defines the learning agent and the possibility of interacting with the process;
  • Process creation defines the dynamic model of the process and the associated interface;
  • Reward creation defines the mathematical expression of the Reward in order to measure the performance when undertaken the proposed task;
  • Agent training is an agent that is trained in order to fulfill the Policy based on the Reward, learning algorithm, and the process.
  • Agent validation is a step in which the performances are evaluated after the training stage;
  • Deploy policy is a step that performs the implementation of the trained agent in the control system (for example, generating the executable code for the embedded system).
An agent used in continuous-type systems is the DDPG, whose improved variant, the TD3 agent, is used in this paper. It is characterized by the fact that it is an actor-critic agent [41] that calculates the long-term maximization of the Reward.
In terms of the training stage, the following stages are completed at each step [38,39]:
  • For the Observation of the current state S, action A = μ ( S ) + N is selected, where N is the stochastic noise from the noise model;
  • Action A is executed, and Reward R and the next Observation S’ are calculated;
  • The experience S , A , R , S is stored;
  • M experiences S i , A i , R i , S i are randomly generated;
  • For S i , which is a terminal state, the value function target yi is set to Ri.
Otherwise, the following relation is calculated [38,41]:
y i = R i + γ min Q k S k , c l i p μ S k θ μ + ε   θ Q k
The value function target is the sum of the experience Reward Ri and the minimum discounted future Reward from the critics.
At each training step, the parameters of each critic are updated, minimizing the following expression:
L k = 1 M i = 1 M y i Q k S i , A i θ Q k 2
At each step, the actor’s parameters are updated, maximizing the Reward:
θ μ J = 1 M i = 1 M G a i G μ i
where Gai, Gμi, and A are given by the following expressions:
G a i = A min Q k S i , A θ Q
G μ i = θ μ μ S i θ μ
A = μ S i θ μ
Furthermore, the parametric updating is performed, for a chosen smoothing coefficient τ, in the next form:
θ Q k = τ θ Q k + 1 τ θ Q k
θ μ = τ θ μ + 1 τ θ μ

3. Correction of the Control Signals for PMSM Based on FOC Strategy Using the RL Agent

The equations of a PMSM are usually expressed using a d-q reference frame using the following relations [18]:
d i d d t = R s L d i d + L q L d n p ω i q + 1 L d u d d i q d t = R s L q i q L d L q n p ω i d λ 0 L q n p ω + 1 L q u q d ω d t = 3 2 n p J λ 0 i q + L d L q i d i q 1 J T L B J ω d θ e d t = n p ω
where: ud is the voltage on the d-axis, uq is the voltage the q-axis, Rd is the resistance on the d-axis, Rq is the resistance on the q-axis, Ld is the inductance on the d-axis, Lq is the inductance on the q-axis, id is the current on the d-axis, iq is the current on the q-axis, ω is the PMSM rotor speed, B is the viscous friction coefficient, J is the rotor inertia, np is the number of pole pairs, TL is the load torque, λ0 is the flux linkage, and θe is the electrical angle of the PMSM rotor.
The classic control structure of the PMSM is the FOC-type structure, and is presented in Figure 2. In this paper, the RL-TD3 agent is used, which will learn the behavior of the PMSM control system. In turn, this system supplies the correction signals for the three control inputs of the cascade control system (iqref, udref, uqref) after the training phase, so that the improved control system will have superior performance.
In the first step, the deep neural network is created, which is characterized by two inputs (Observation and Action) and one output. The code sequence of the program developed in the MATLAB environment for the creation of the neural network is presented in Figure 3, and its graphic representation is shown in Figure 4.
Training of the TD3 agent for the PMSM control uses a maximum of 200 episodes, the number of steps per episode is 100, and the sampling step of the agent is 10−4 s. The agent training stage stops when the cumulative average Reward is greater than −190 for a period of 100 consecutive episodes, or after the 200 training episodes initially set have passed. To improve the learning performance during training, a Gaussian noise overlaps the signals received and transmitted by the agent.

3.1. Reinforcement Learning Agent for the Correction of the Outer Speed Control Loop

The block diagram of the implementation in MATLAB/Simulink for the PMSM control of the outer loop (which controls the speed of PMSM) based on the RL-TD3 agent is presented in Figure 5. Figure 6 shows the RL block structure. In this case, the correction signals of TD3 agent will be added to the control signal iqref. The Observation consists of signals ω and ωerror.
The Reward at each step for this case is expressed as follows:
r 1 = Q 1 ω e r r o r 2 + R j u t 1 j 2
where: Q1 = 0.5 and R = 0.1.
The training time for this case is 6 h, 23 min, and 27 s. The graphical results for the training stage of the RL-TD3 agent for the correction of iqref are shown in Figure 7.

3.2. Reinforcement Learning Agent for the Correction of the Inner Currents Control Loop

The block diagram of the implementation in MATLAB/Simulink for the PMSM control of the inner loop (which controls the id and iq currents) based on the RL-TD3 agent is presented in Figure 8. After the learning stage, the RL-TD3 agent will supply correction signals for the control signals ud and uq. Figure 9 shows the RL block structure. The Observation consists of signals id, iq, iderror, and iqerror.
The Reward at each step for this case is expressed as follows:
r 1 = Q 1 i d e r r o r 2 + Q 2 i q e r r o r 2 + R j u t 1 j 2
where: Q1 = Q2 = 0.5, R = 0.1, and u t 1 j represents the actions from the previous time step.
The training time for this case is 6 h, 8 min, and 2 s. The graphical results for the training stage of the RL-TD3 agent for the correction of udref and uqref are shown in Figure 10.

3.3. Reinforcement Learning Agent for the Correction of the Outer Speed Control Loop and Inner Currents Control Loop

The block diagram of the implementation in MATLAB/Simulink for the correction of the inner current control loop and outer speed control loop based on the RL-TD3 agent is presented in Figure 11. Figure 12 shows the RL block structure. In this case, the correction signals of RL-TD3 agent will be supplied to the control signals ud and uq, and also to the iqref signal. The Observation consists of signals ω, ωerror, id, iq, iderror, and iqerror.
The Reward at each step for this case is expressed as follows:
r 1 = Q 1 ω e r r o r 2 + Q 2 i d e r r o r 2 + Q 3 i q e r r o r 2 + R j u t 1 j 2
where: Q1 = Q2 = Q3 = 0.5 and R = 0.1.
The training time for this case is 7 h, 59 min, and 35 s. The graphical results for the training stage of the RL-TD3 agent for the correction of ud, uq, iqref are shown in Figure 13.

4. Correction of the Control Signals for PMSM—FOC Strategy Based on SMC and Synergetic Control Using the Reinforcement Learning Agent

Based on the classic FOC-type structure in which the usual PI or hysteresis controllers are replaced with SMC or synergetic controllers, respectively, Figure 14 presents the block diagram for the SMC and synergetic control of the PMSM based on the reinforcement learning. The following subsection presents the PMSM control in which an SMC-type controller is proposed for the outer speed control loop, and a synergetic-type controller is proposed for the inner current control loop.

4.1. SMC and Synergetic for PMSM Control

4.1.1. SMC Speed Controller Description and MATLAB/Simulink Implementation

Based on Equation (9), rewritten in the form of Equation (13), to obtain the SMC control law, the notations are expressed in Equations (14) and (15) in which the state variables x1 and x2 are defined [17,18,42,43].
i ˙ d i ˙ q ω ˙ = R s L n p ω 0 n p ω R s L n p λ 0 L 0 K t J B J i d i q ω + u d L u q L T L J
x 1 = ω r e f ω
x 2 = x ˙ 1 = ω r e f ω d t = ω ˙
It can be noted that variable x1 represents the tracking error of the PMSM speed and x2 represents its derivative.
The sliding surface S and its derivative are defined in Equations (16) and (17), respectively.
S = c x 1 + x 2
S ˙ = c x 2 + x ˙ 2 = c x 2 D i ˙ q
where: c is a positive adjustable parameter and D = 3 2 n p λ 0 J .
In Equation (18), the condition of the evolution on surface S is imposed as follows:
S ˙ = ε sgn ( S ) q S ,       ε , q > 0
where: ε and q represent the positive adjustable parameters; sgn() represents the signum function.
By using the function defined in Equation (19), a reduction in the chattering effect is obtained instead of the function sgn [43]:
H ( x ) = 2 1 + e a ( x c ) 1
For a = 4 and c = 0, H ∈ [−1 1], the evolution of the function H from −1 to 1 is smoothed, thus reducing the chattering effect.
After some calculations, the SMC-type controller output can be expressed as follows:
i q r e f ( t ) = 1 D 0 t c x 2 + ε H ( S ) + q S d t
To demonstrate the stability, a Lyapunov function is chosen of the form V = 1 2 S 2 . By calculating the derivative of function V, we obtain its negativity according to Equation (21):
V ˙ = S S ˙ = S ε H ( S ) q S = ε H ( S ) q S 2
The block diagram of the MATLAB/Simulink implementation for the SMC law control is presented in Figure 15.

4.1.2. Synergetic Currents Controller Description and MATLAB/Simulink Implementation

By adding a macro-variable as a function of the states of the system in Equation (22) and by imposing an evolution of its dynamics in the form of Equation (23), it is noted that the synergetic control law design is a generalization of SMC control law design [19,42,43]:
Ψ = Ψ ( x , t )
T Ψ ˙ + Ψ = 0 ,   T > 0
where: T represents the rate of convergence of the system evolution to achieve the desired manifold.
The differentiation of the macro-variable Ψ can be written by Equation (24):
Ψ ˙ = d Ψ d x x ˙ ,  
The general form of the control law for the PMSM is obtained by combining Equations (13), (23) and (24):
u = u x , Ψ ( x , t ) , T , t
Following [20,42], superior performances can be obtained for static and dynamic regimes by selecting threshold values according to Equations (26) and (27):
ω a c c = ω r e f k q i q max i q r e f
ω d e c = ω r e f k q i q max i q r e f
Based on these threshold values that practically limit the acceleration and deceleration regimes, the macro-variable can be defined on the q-axis Ψq. Therefore, for ω > ω a c c and ω < ω d e c the following relation is defined:
Ψ q = ω ( t ) ω r e f + k q i q ( t ) i q r e f
For ω ω a c c , the corresponding macro-variable is defined as in Equation (29), and for ω ω d e c the corresponding macro-variable is defined as in Equation (30).
Ψ q = i q ( t ) i q max + k i q 0 t i q ( t ) i q max d t
Ψ q = i q ( t ) + i q max + k i q 0 t i q ( t ) + i q max d t
where: iqmax represents the maximum admissible current on the q-axis, and kq represents a value that is dynamically adjusted as a function of the PMSM rotor speed error.
Based on these equations, the control law for the q-axis can be expressed according to the acceleration and deceleration subdomains presented above, as follows: Equation (31) for ω a c c < ω < ω d e c , uq; Equation (33) for ω ω a c c , uq; and Equation (32) for ω ω d e c , uq:
u q ( t ) = R s i q + n p ω L i d + λ 0 + L T q i q r e f i q + L T q k q ω r e f ω + L J k q K t i q + B ω + T L
u q ( t ) = R s i q + n p ω L i d + λ 0 + L T q i q max i q + k i q L i q max i q k i q L T q 0 t i q i q max d t
u q ( t ) = R s i q + n p ω L i d + λ 0 + L T q i q max i q + k i q L i q max i q k i q L T q 0 t i q + i q max d t
The block diagram of the MATLAB/Simulink implementation for the control law on the q-axis corresponding to Equations (31)–(33) is presented in Figure 16, Figure 17 and Figure 18, respectively.
The macro-variable on the d-axis Ψd can be expressed as in Equation (34):
Ψ d = i d ( t ) i d r e f + k i d 0 t i d ( t ) i d r e f d t
After some calculus, the control law for the d-axis can be expressed as follows [20,43]:
u d ( t ) = R s i d n p ω L i q L T d i d i d r e f k i d L i d i d r e f k i d L T d 0 t i d i d r e f d t
The block diagram of the MATLAB/Simulink implementation for the control law on the d-axis corresponding to Equation (35) is presented in Figure 19.

4.1.3. Speed Observer

In the α-β frame, using the PMSM equations derived from Equation (13) after applying the inverse Park transform, Equations (36)–(38) present the expressions of currents iα and iβ, their derivatives, and the back-EMF eα and eβ [18,25,43]:
i α = i d cos ( θ e ) i q sin ( θ e ) i β = i d sin ( θ e ) + i q cos ( θ e )
e α = d λ α d t = λ 0 ω e sin θ e e β = d λ β d t = λ 0 ω e cos θ e
d i α d t = R s L i α 1 L e α + 1 L u α d i β d t = R s L i β 1 L e β + 1 L u β
The equations of the SMO-type observer can be written as follows [18,25,43]:
d i ^ α d t = R s L i ^ α + 1 L u α 1 L k H ( i ^ α i α ) d i ^ β d t = R s L i ^ β + 1 L u β 1 L k H ( i ^ β i β )
where: k is the observer gain, and H represents the sigmoid type function described in Equation (19).
The sliding vector is chosen as follows:
S n = [ S α   S β ] T = [ i ^ α i α   i ^ β i β ] T = [ i ¯ α   i ¯ β ] T
Moreover, a Lyapunov function is chosen in the form of Equation (41) [25,43].
V = 1 2 S n T S n = 1 2 S α 2 + S β 2
For the calculation of the derivative of this function, the current error system is defined as follows:
i ¯ ˙ α = i ^ ˙ α i ˙ α = R s L i ¯ α + 1 L e α 1 L k H ( i ¯ α ) i ¯ ˙ β = i ^ ˙ β i ˙ β = R s L i ¯ β + 1 L e β 1 L k H ( i ¯ β )
Based on these, Equation (43) is obtained.
V ˙ = R s L i ¯ α 2 + i ¯ β 2 + 1 L e α k i ¯ α H ( i ¯ α ) + e β k i ¯ β H ( i ¯ β ) < 0
By choosing the observer gain of the form k max e α , e β , the stability condition of the speed observer is obtained: V ˙ < 0 .
In addition, the following relation can be written:
[ S ˙ α   S ˙ β ] T = [ S α   S β ] T [ 0   0 ]
Using Equations (43) and (44) the estimations for eα and eβ are achieved:
e ^ α = k H ( i ¯ α ) = λ 0 ω ^ e sin θ e e ^ β = k H ( i ¯ β ) = λ 0 ω ^ e cos θ e
Equations (46) and (47) show the estimations of the PMSM rotor speed and positions.
ω ^ e = e ^ α 2 + e ^ β 2 λ 0
θ ^ e ( t ) = t 0 t ω ^ e ( t ) d t + θ 0
where: θ0 is the initial electrical position of the PMSM rotor.
The block diagram of the MATLAB/Simulink implementation for the PMSM rotor speed and position estimations based on the SMO-type observer is presented in Figure 20.
The block diagram of the MATLAB/Simulink implementation for the SMC and synergetic sensorless control of the PMSM (developed based on the elements presented in this subsection) is presented in Figure 21. For the implementation of the SMC-type controller described in Section 4.1.1, the parameters ε = 300, q = 200, and c = 100 were selected, and for the synergetic type controller described in Section 4.1.2, the parameters kiq = 10,000, kq = 10,000, iqmax = 50, Td = 3, Tq = 3, and kid = 10,000 were selected.

4.2. Reinforcement Learning Agent for the Correction of the Outer Speed Control Loop Using SMC and Synergetic Control

By analogy with the aspects presented in Section 3, this section continues with the presentation of the way in which the performance of the PMSM control system can be improved using the RL, even if the controllers used in the FOC-type control strategy are complex SMC and synergetic controllers. A maximum number of 200 training episodes was chosen for the RL-TD3 agent, and for each episode the number of steps is 100, with a sampling period of 10−4 s. The RL-TD3 agent training stage is stopped when the cumulative average Reward is greater than −190 for a period of 100 consecutive episodes, or after 200 training episodes have run. A Gaussian noise is overlapped on the original signals, received and transmitted by the agent, to improve the learning performance.
The block diagram of the implementation in MATLAB/Simulink for the PMSM control using SMC and synergetic controllers, resulting in improved performance using the RL-TD3 agent in the outer control loop, is presented in Figure 22.
The correction signals of the RL-TD3 agent will be added to the control signal iqref, the RL block structure is similar to that of Figure 6, and the Reward is given by Equation (10). The training time for this case is 10 h, 17 min, and 31 s. The graphical results for the training stage of RL-TD3 agent for the correction of iqref are shown in Figure 23.

4.3. Reinforcement Learning Agent for the Correction of the Inner Currents Control Loop Using SMC and Synergetic Control

The block diagram of the implementation in MATLAB/Simulink for the PMSM control using SMC and synergetic controllers, resulting in improved performance using the RL-TD3 agent in the inner control loop, is presented in Figure 24. The correction signals of the RL-TD3 agent will be added to the control signals udref and uqref, the RL block structure is similar to that of Figure 9, and the Reward is given by Equation (11).
The training time for this case is 10 h, 29 min, and 14 s. The graphical results for the training stage of the RL-TD3 agent for the correction of udref and uqref are shown in Figure 25.

4.4. Reinforcement Learning Agent for the Correction of the Outer Speed Control Loop and Inner Currents Control Loop Using SMC and Synergetic Control

The block diagram of the implementation in MATLAB/Simulink for the PMSM control using SMC and synergetic controllers, resulting in the improved performance using the RL-TD3 agent in the outer and inner control loops, is presented in Figure 26. The correction signals of the RL-TD3 agent will be added to the control signals iqref, udref, and uqref, the RL block structure is similar to that of Figure 12, and the Reward is given by Equation (12).
The training time for this case is 11 h, 13 min, and 54 s. The graphical results for the training stage of the RL-TD3 agent for the correction of iqref, udref, and uqref are shown in Figure 27.

5. Numerical Simulations

For a PMSM with the nominal parameters presented in Table 1, using MATLAB/Simulink, Figure 28 shows the time evolution of the signals of interest of the PMSM control system. For the variation in the reference signal of the PMSM rotor speed from 800 to 1200 rpm at 0.5 s and a load torque of 0.5 Nm, a zero steady-state error and a response time of 30 ms are noted.
As a result of the creation of RL-TD3 agents, their training, and the numerical simulations related to the cases in Section 3.1, Section 3.2 and Section 3.3 of the Section 3, Figure 29, Figure 30 and Figure 31 show the corresponding results obtained.
In the case of the RL agent for the correction of the outer loop for speed control (Figure 29), the stator currents remain identical to those in the classic case of the FOC-type control structure, but the response time improves very little, in the sense that it decreases by approximately 1 ms.
In the case of the RL-TD3 agent for the correction of the inner current control loop (Figure 30), an improvement is noted in the response of the control system, i.e., a shorter response time of 25 ms, but the stator currents at start-up increase by 50%. In this case, the current decreases to the usual values by decreasing the hysteresis of the current controllers, but a value of the response time of approximately 28 ms is obtained.
In the case of the RL-TD3 agent for the correction of the outer speed control loop and the inner current control loop (Figure 31), a very good global response is obtained, namely the response time decreases to 24 ms under the conditions where the stator currents remain similar to those in the classic case of the FOC-type control structure.
The efficiency of using the RL is thereby demonstrated by improving the performance of the PMSM control system when using the FOC-type strategy with PI-type controllers for speed control and hysteresis ON/OFF for current control. Depending on the application, among the presented variants, we choose the one that achieves an acceptable compromise between the decrease in the response time and the increase in the stator currents.
Based on articles [42,43], in which the FOC-type structure with SMC-type controllers was chosen for speed control and synergetic-type controllers were chosen for current control, and peak performance was obtained on benchmarks (including for the PMSM control), the following simulations validate that these performances can be further improved by using the RL.
Thus, in Figure 32, for a FOC-type structure based on the SMC and synergetic controllers, the response of the PMSM control system is presented under the same conditions in which the FOC-type structure is based on PI and hysteresis ON/OFF controllers. Very good control performance is noted, with a response time of 14.1 ms.
Moreover, Figure 33, Figure 34 and Figure 35 show the response of the PMSM control system under the same conditions as those presented above, in which the control signal is supplemented with the output of an RL-TD3 agent for the speed control loop, for the current control loop, and for both control loops.
The validation of the robustness of the PMSM control system based on SMC and synergetic controllers, and the three variants of control signal correction implemented by means of the RL-TD3 agent, were achieved by modifying the load torque TL and the combined inertia of the PMSM rotor and load J, which are considered to be disturbances. Thus, numerical simulations were performed for a value of TL = 2 Nm, to which a 0.2 Nm magnitude uniformly distributed noise and a 50% J parameter increase were added. Following the numerical simulations, Figure 36, Figure 37, Figure 38 and Figure 39 shows the response of the PMSM control system under parametric changes. It can be noted that the PMSM control system retains its previous performance, noting that the response time increases by about 0.4 ms for each variant of the control system.
Figure 40 shows the comparative response of the PMSM control system for an 800 rpm reference speed step, a load torque of 0.5 Nm for the of PMSM control systems based on SMC and synergetic controllers, and three variants of this type of control system using an RL-TD3 agent for correction of the command signals (outer and inner control loop).
Table 2 shows the comparative performance of these types of PMSM control system variants, i.e., the response time and the ripple of the PMSM rotor speed error signal given by Equation (48). In all these cases the overshooting is zero, and the steady-state error is less than 0.1%.
ω r i p = 1 N i = 1 N ω ( i ) ω r e f ( i ) 2
where: N represents the number of samples, ω represents the PMSM rotor speed, and ωref represents the reference speed.
A previous study [43] described in detail the implementation of real-time control algorithms (such as SMC and synergetic controllers) in an S32K144 development kit containing an S32K144 evaluation board (S32K144EVB-Q100), a DEVKIT-MOTORGD board based on a SMARTMOS GD3000 pre-driver, and a Linix 45ZWN24-40 PMSM type. The controller of the development platform was an S32K144 MCUm, which is a 32-bit Cortex M4F type having a time base of 112 MHz with 512 KB of flash memory and 54 KB of RAM. The presentation in real time of the improvement resulting from the current article in the PMSM control system on the same benchmark was omitted, because it was considered less important than the presentation of the RL-TD3 agent control structures and the advantages they provided.

6. Conclusions

This paper presents the FOC-type control structure of a PMSM, which is improved in terms of performance by using a RL technique. Thus, the comparative results are presented for the case where the RL-TD3 agent is properly trained and provides correction signals that are added to the control signals ud, uq, and iqref. The FOC-type control structure for the PMSM control based on an SMC speed controller and synergetic current controller is also presented. To improve the performance of the PMSM control system without using controllers having a more complicated mathematical description, the advantages provided by the RL on process control can also be used. This improvement is obtained using the correction signals provided by a trained RL-TD3 agent, which is added to the control signals ud, uq, and iqref. A speed observer is also implemented for estimating the PMSM rotor speed. The parametric robustness of the proposed PMSM control system is proved by very good control performances achieved even when the uniformly distributed noise is added to the load torque TL, and under high variations in the load torque TL and the moment of inertia J. Numerical simulations are used to prove the superiority of the control system that uses the RL-TD3 agent.

Author Contributions

Conceptualization, M.N. and C.-I.N.; Data curation, M.N., C.-I.N. and D.S.; Formal analysis, M.N., C.-I.N. and D.S.; Funding acquisition, M.N. and D.S.; Investigation, M.N., C.-I.N. and D.S.; Methodology, M.N., C.-I.N. and D.S.; Project administration, M.N. and D.S.; Resources, M.N. and D.S.; Software, M.N. and C.-I.N.; Supervision, M.N. and D.S.; Validation, M.N. and D.S.; Visualization, M.N., C.-I.N. and D.S.; Writing—original draft, M.N., C.-I.N. and D.S.; Writing—review and editing, M.N., C.-I.N. and D.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by European Regional Development Fund Competitiveness Operational Program, project TISIPRO, ID: P_40_416/105736, 2016–2021 and with funds from the Ministry of Research and Innovation—Romania as part of the NUCLEU Program: PN 19 38 01 03.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Eriksson, S. Design of Permanent-Magnet Linear Generators with Constant-Torque-Angle Control for Wave Power. Energies 2019, 12, 1312. [Google Scholar] [CrossRef] [Green Version]
  2. Ouyang, P.R.; Tang, J.; Pano, V. Position domain nonlinear PD control for contour tracking of robotic manipulator. Robot. Comput. Integr. Manuf. 2018, 51, 14–24. [Google Scholar] [CrossRef]
  3. Baek, S.W.; Lee, S.W. Design Optimization and Experimental Verification of Permanent Magnet Synchronous Motor Used in Electric Compressors in Electric Vehicles. Appl. Sci. 2020, 10, 3235. [Google Scholar] [CrossRef]
  4. Amin, F.; Sulaiman, E.B.; Utomo, W.M.; Soomro, H.A.; Jenal, M.; Kumar, R. Modelling and Simulation of Field Oriented Control based Permanent Magnet Synchronous Motor Drive System. Indones. J. Electr. Eng. Comput. Sci. 2017, 6, 387. [Google Scholar] [CrossRef]
  5. Mohd Zaihidee, F.; Mekhilef, S.; Mubin, M. Robust Speed Control of PMSM Using Sliding Mode Control (SMC)—A Review. Energies 2019, 12, 1669. [Google Scholar] [CrossRef] [Green Version]
  6. Maleki, N.; Pahlavani, M.R.A.; Soltani, I. A Detailed Comparison between FOC and DTC Methods of a Permanent Magnet Synchronous Motor Drive. J. Electr. Electron. Eng. 2015, 3, 92–100. [Google Scholar] [CrossRef] [Green Version]
  7. Sakunthala, S.; Kiranmayi, R.; Mandadi, P.N. A Review on Speed Control of Permanent Magnet Synchronous Motor Drive Using Different Control Techniques. In Proceedings of the International Conference on Power, Energy, Control and Transmission Systems (ICPECTS), Chennai, China, 22–23 February 2018; pp. 97–102. [Google Scholar]
  8. Su, G.; Wang, P.; Guo, Y.; Cheng, G.; Wang, S.; Zhao, D. Multiparameter Identification of Permanent Magnet Synchronous Motor Based on Model Reference Adaptive System—Simulated Annealing Particle Swarm Optimization Algorithm. Electronics 2022, 11, 159. [Google Scholar] [CrossRef]
  9. Nicola, M.; Nicola, C.-I. Sensorless Control for PMSM Using Model Reference Adaptive Control and back -EMF Sliding Mode Observer. In Proceedings of the International Conference on Electromechanical and Energy Systems (SIELMEN), Craiova, Romania, 9–10 October 2019; pp. 1–7. [Google Scholar]
  10. Wang, S.; Zhao, J.; Liu, T.; Hua, M. Adaptive Robust Control System for Axial Flux Permanent Magnet Synchronous Motor of Electric Medium Bus Based on Torque Optimal Distribution Method. Energies 2019, 12, 4681. [Google Scholar] [CrossRef] [Green Version]
  11. Bao, G.; Qi, W.; He, T. Direct Torque Control of PMSM with Modified Finite Set Model Predictive Control. Energies 2020, 13, 234. [Google Scholar] [CrossRef] [Green Version]
  12. Tang, M.; Zhuang, S. On Speed Control of a Permanent Magnet Synchronous Motor with Current Predictive Compensation. Energies 2019, 12, 65. [Google Scholar] [CrossRef] [Green Version]
  13. Zhang, G.; Chen, C.; Gu, X.; Wang, Z.; Li, X. An Improved Model Predictive Torque Control for a Two-Level Inverter Fed Interior Permanent Magnet Synchronous Motor. Electronics 2019, 8, 769. [Google Scholar] [CrossRef] [Green Version]
  14. Liu, X.; Zhang, Q. Robust Current Predictive Control-Based Equivalent Input Disturbance Approach for PMSM Drive. Electronics 2019, 8, 1034. [Google Scholar] [CrossRef] [Green Version]
  15. Ma, Y.; Li, Y. Active Disturbance Compensation Based Robust Control for Speed Regulation System of Permanent Magnet Synchronous Motor. Appl. Sci. 2020, 10, 709. [Google Scholar] [CrossRef] [Green Version]
  16. Lin, C.-H. Permanent-Magnet Synchronous Motor Drive System Using Backstepping Control with Three Adaptive Rules and Revised Recurring Sieved Pollaczek Polynomials Neural Network with Reformed Grey Wolf Optimization and Recouped Controller. Energies 2020, 13, 5870. [Google Scholar] [CrossRef]
  17. Merabet, A. Cascade Second Order Sliding Mode Control for Permanent Magnet Synchronous Motor Drive. Electronics 2019, 8, 1508. [Google Scholar] [CrossRef] [Green Version]
  18. Utkin, V.; Guldner, J.; Shi, J. Automation and Control Engineering. In Sliding Mode Control in Electromechanical Systems, 2nd ed.; CRC Press: Boca Raton, FL, USA, 2009; pp. 223–271. [Google Scholar]
  19. Bounasla, N.; Hemsas, K.E.; Mellah, H. Synergetic and sliding mode controls of a PMSM: A comparative study. J. Electr. Electron. Eng. 2015, 3, 22–26. [Google Scholar] [CrossRef] [Green Version]
  20. Bogani, T.; Lidozzi, A.; Solero, L.; Di Napoli, A. Synergetic Control of PMSM Drives for High Dynamic Applications. In Proceedings of the IEEE International Conference on Electric Machines and Drives, San Antonio, TX, USA, 15 May 2005; pp. 710–717. [Google Scholar]
  21. Chang, Y.-C.; Tsai, C.-T.; Lu, Y.-L. Current Control of the Permanent-Magnet Synchronous Generator Using Interval Type-2 T-S Fuzzy Systems. Energies 2019, 12, 2953. [Google Scholar] [CrossRef] [Green Version]
  22. Nicola, M.; Nicola, C.-I. Improved Performance of Sensorless Control for PMSM Based on Neuro-fuzzy Speed Controller. In Proceedings of the 9th International Conference on Modern Power Systems (MPS), Cluj-Napoca, Romania, 16–17 June 2021; pp. 1–6. [Google Scholar]
  23. Hoai, H.-K.; Chen, S.-C.; Than, H. Realization of the Sensorless Permanent Magnet Synchronous Motor Drive Control System with an Intelligent Controller. Electronics 2020, 9, 365. [Google Scholar] [CrossRef] [Green Version]
  24. Nicola, M.; Nicola, C.-I.; Sacerdoțianu, D. Sensorless Control of PMSM using DTC Strategy Based on Multiple ANN and Load Torque Observer. In Proceedings of the 21st International Symposium on Electrical Apparatus & Technologies (SIELA), Bourgas, Bulgaria, 3–6 June 2020; pp. 1–6. [Google Scholar]
  25. Wang, M.-S.; Tsai, T.-M. Sliding Mode and Neural Network Control of Sensorless PMSM Controlled System for Power Consumption and Performance Improvement. Energies 2017, 10, 1780. [Google Scholar] [CrossRef] [Green Version]
  26. Nicola, M.; Nicola, C.-I. Tuning of PI Speed Controller for PMSM Control System Using Computational Intelligence. In Proceedings of the 21st International Symposium on Power Electronics (Ee), Novi Sad, Serbia, 27–30 October 2021; pp. 1–6. [Google Scholar]
  27. Zhu, Y.; Tao, B.; Xiao, M.; Yang, G.; Zhang, X.; Lu, K. Luenberger Position Observer Based on Deadbeat-Current Predictive Control for Sensorless PMSM. Electronics 2020, 9, 1325. [Google Scholar] [CrossRef]
  28. Nicola, M.; Nicola, C.-I.; Sacerdotianu, D. Sensorless Control of PMSM using DTC Strategy Based on PI-ILC Law and MRAS Observer. In Proceedings of the 2020 International Conference on Development and Application Systems (DAS), Suceava, Romania, 21–23 May 2020; pp. 38–43. [Google Scholar]
  29. Ye, M.; Shi, T.; Wang, H.; Li, X.; Xiai, C. Sensorless-MTPA Control of Permanent Magnet Synchronous Motor Based on an Adaptive Sliding Mode Observer. Energies 2019, 12, 3773. [Google Scholar] [CrossRef] [Green Version]
  30. Urbanski, K.; Janiszewski, D. Sensorless Control of the Permanent Magnet Synchronous Motor. Sensors 2019, 19, 3546. [Google Scholar] [CrossRef] [Green Version]
  31. Kamel, T.; Abdelkader, D.; Said, B.; Padmanaban, S.; Iqbal, A. Extended Kalman Filter Based Sliding Mode Control of Parallel-Connected Two Five-Phase PMSM Drive System. Electronics 2018, 7, 14. [Google Scholar] [CrossRef] [Green Version]
  32. Morimoto, R.; Nishikawa, S.; Niiyama, R.; Kuniyoshi, Y. Model-Free Reinforcement Learning with Ensemble for a Soft Continuum Robot Arm. In Proceedings of the IEEE 4th International Conference on Soft Robotics (RoboSoft), New Haven, CT, USA, 12–16 April 2021; pp. 141–148. [Google Scholar]
  33. Murti, F.W.; Ali, S.; Latva-aho, M. Deep Reinforcement Based Optimization of Function Splitting in Virtualized Radio Access Networks. In Proceedings of the IEEE International Conference on Communications Workshops (ICC Workshops), Montreal, QC, Canada, 14–23 June 2021; pp. 1–6. [Google Scholar]
  34. Song, S.; Foss, L.; Dietz, A. Deep Reinforcement Learning for Permanent Magnet Synchronous Motor Speed Control Systems. Neural Comput. Appl. 2020, 33, 5409–5418. [Google Scholar] [CrossRef]
  35. Schindler, T.; Ali, S.; Latva-aho, M. Comparison of Reinforcement Learning Algorithms for Speed Ripple Reduction of Permanent Magnet Synchronous Motor. In Proceedings of the IKMT 2019—Innovative Small Drives and Micro-Motor Systems; 12. ETG/GMM-Symposium, Wuerzburg, Germany, 10–11 September 2019; pp. 1–6. [Google Scholar]
  36. Brandimarte, P. Approximate Dynamic Programming and Reinforcement Learning for Continuous States. In From Shortest Paths to Reinforcement Learning: A MATLAB-Based Tutorial on Dynamic Programming; Springer Nature: Cham, Switzerland, 2021; pp. 185–204. [Google Scholar]
  37. Beale, M.; Hagan, M.; Demuth, H. Deep Learning Toolbox™ Getting Started Guide, 14th ed.; MathWorks, Inc.: Natick, MA, USA, 2020. [Google Scholar]
  38. Nicola, M.; Nicola, C.-I. Improvement of PMSM Control Using Reinforcement Learning Deep Deterministic Policy Gradient Agent. In Proceedings of the 21st the International Symposium on Power Electronics (Ee), Novi Sad, Serbia, 27–30 October 2021; pp. 1–6. [Google Scholar]
  39. Nicola, M.; Nicola, C.-I. Improved Performance for PMSM Control System Based on LQR Controller and Computational Intelligence. In Proceedings of the International Conference on Electrical, Computer and Energy Technologies (ICECET), Cape Town, South Africa, 9–10 December 2021; pp. 1–6. [Google Scholar]
  40. Nicola, M.; Nicola, C.-I. Improved Performance for PMSM Control System Based on Feedback Linearization and Computational Intelligence. In Proceedings of the International Conference on Electrical, Computer and Energy Technologies (ICECET), Cape Town, South Africa, 9–10 December 2021; pp. 1–6. [Google Scholar]
  41. MathWorks—Reinforcement Learning Toolbox™ User’s Guide. Available online: https://www.mathworks.com/help/reinforcement-learning/getting-started-with-reinforcement-learning-toolbox.html?s_tid=CRUX_lftnav (accessed on 4 November 2021).
  42. Nicola, M.; Nicola, C.-I. Fractional-Order Control of Grid-Connected Photovoltaic System Based on Synergetic and Sliding Mode Controllers. Energies 2021, 14, 510. [Google Scholar] [CrossRef]
  43. Nicola, M.; Nicola, C.-I. Sensorless Fractional Order Control of PMSM Based on Synergetic and Sliding Mode Controllers. Electronics 2020, 9, 1494. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram for a RL scenario.
Figure 1. Schematic diagram for a RL scenario.
Energies 15 02208 g001
Figure 2. Block diagram for FOC-type control of the PMSM based on PI-type controllers using RL.
Figure 2. Block diagram for FOC-type control of the PMSM based on PI-type controllers using RL.
Energies 15 02208 g002
Figure 3. Creation of the deep neural network—MATLAB syntax code example.
Figure 3. Creation of the deep neural network—MATLAB syntax code example.
Energies 15 02208 g003
Figure 4. Representation of the deep neural network for the RL-TD3 agent.
Figure 4. Representation of the deep neural network for the RL-TD3 agent.
Energies 15 02208 g004
Figure 5. Block diagram of the MATLAB/Simulink implementation for the PMSM control based on PI-type controllers using the RL-TD3 agent for the correction of iqref.
Figure 5. Block diagram of the MATLAB/Simulink implementation for the PMSM control based on PI-type controllers using the RL-TD3 agent for the correction of iqref.
Energies 15 02208 g005
Figure 6. MATLAB/Simulink implementation of RL-TD3 agent for the correction of iqref.
Figure 6. MATLAB/Simulink implementation of RL-TD3 agent for the correction of iqref.
Energies 15 02208 g006
Figure 7. Training stage of the RL-TD3 agent for the correction of iqref.
Figure 7. Training stage of the RL-TD3 agent for the correction of iqref.
Energies 15 02208 g007
Figure 8. Block diagram of the MATLAB/Simulink implementation for the PMSM control based on PI-type controllers using the RL-TD3 agent for the correction of udref and uqref.
Figure 8. Block diagram of the MATLAB/Simulink implementation for the PMSM control based on PI-type controllers using the RL-TD3 agent for the correction of udref and uqref.
Energies 15 02208 g008
Figure 9. MATLAB/Simulink implementation of the RL-TD3 agent for the correction of udref and uqref.
Figure 9. MATLAB/Simulink implementation of the RL-TD3 agent for the correction of udref and uqref.
Energies 15 02208 g009
Figure 10. Training stage of the RL-TD3 agent for the correction of udref and uqref.
Figure 10. Training stage of the RL-TD3 agent for the correction of udref and uqref.
Energies 15 02208 g010
Figure 11. Block diagram of MATLAB/Simulink implementation for the PMSM control based on PI-type controllers using the RL-TD3 agent for the correction of udref, uqref, and iqref.
Figure 11. Block diagram of MATLAB/Simulink implementation for the PMSM control based on PI-type controllers using the RL-TD3 agent for the correction of udref, uqref, and iqref.
Energies 15 02208 g011
Figure 12. MATLAB/Simulink implementation of the RL- TD3 agent for the correction of udref, uqref, and iqref.
Figure 12. MATLAB/Simulink implementation of the RL- TD3 agent for the correction of udref, uqref, and iqref.
Energies 15 02208 g012
Figure 13. Training stage of the RL-TD3 agent for the correction of udref, uqref, and iqref.
Figure 13. Training stage of the RL-TD3 agent for the correction of udref, uqref, and iqref.
Energies 15 02208 g013
Figure 14. Block diagram for the SMC and synergetic control of PMSM based on RL.
Figure 14. Block diagram for the SMC and synergetic control of PMSM based on RL.
Energies 15 02208 g014
Figure 15. Block diagram of the MATLAB/Simulink implementation for the SMC speed controller.
Figure 15. Block diagram of the MATLAB/Simulink implementation for the SMC speed controller.
Energies 15 02208 g015
Figure 16. Block diagram of the MATLAB/Simulink implementation for the synergetic current controller of the uq command—Equation (31).
Figure 16. Block diagram of the MATLAB/Simulink implementation for the synergetic current controller of the uq command—Equation (31).
Energies 15 02208 g016
Figure 17. Block diagram of the MATLAB/Simulink implementation for the synergetic current controller of the uq command—Equation (32).
Figure 17. Block diagram of the MATLAB/Simulink implementation for the synergetic current controller of the uq command—Equation (32).
Energies 15 02208 g017
Figure 18. Block diagram of the MATLAB/Simulink implementation for the synergetic current controller of the uq command—Equation (33).
Figure 18. Block diagram of the MATLAB/Simulink implementation for the synergetic current controller of the uq command—Equation (33).
Energies 15 02208 g018
Figure 19. Block diagram of the MATLAB/Simulink implementation for the synergetic current controller of the ud command—Equation (35).
Figure 19. Block diagram of the MATLAB/Simulink implementation for the synergetic current controller of the ud command—Equation (35).
Energies 15 02208 g019
Figure 20. Block diagram of the MATLAB/Simulink implementation for PMSM rotor speed and position estimations.
Figure 20. Block diagram of the MATLAB/Simulink implementation for PMSM rotor speed and position estimations.
Energies 15 02208 g020
Figure 21. Block diagram of the MATLAB/Simulink implementation for SMC and synergetic sensorless control of the PMSM.
Figure 21. Block diagram of the MATLAB/Simulink implementation for SMC and synergetic sensorless control of the PMSM.
Energies 15 02208 g021
Figure 22. Block diagram of the MATLAB/Simulink implementation for the PMSM control based on SMC and synergetic controllers using the RL-TD3 agent for the correction of iqref.
Figure 22. Block diagram of the MATLAB/Simulink implementation for the PMSM control based on SMC and synergetic controllers using the RL-TD3 agent for the correction of iqref.
Energies 15 02208 g022
Figure 23. Training stage of the RL-TD3 agent for the correction of iqref.
Figure 23. Training stage of the RL-TD3 agent for the correction of iqref.
Energies 15 02208 g023
Figure 24. Block diagram of the MATLAB/Simulink implementation for the PMSM control based on SMC and synergetic controllers using the RL-TD3 agent for the correction of udref and uqref.
Figure 24. Block diagram of the MATLAB/Simulink implementation for the PMSM control based on SMC and synergetic controllers using the RL-TD3 agent for the correction of udref and uqref.
Energies 15 02208 g024
Figure 25. Training stage of the RL-TD3 agent for the correction of udref and uqref.
Figure 25. Training stage of the RL-TD3 agent for the correction of udref and uqref.
Energies 15 02208 g025
Figure 26. Block diagram of the MATLAB/Simulink implementation for the PMSM control based on the RL-TD3 agent for the correction of udref, uqref, and iqref.
Figure 26. Block diagram of the MATLAB/Simulink implementation for the PMSM control based on the RL-TD3 agent for the correction of udref, uqref, and iqref.
Energies 15 02208 g026
Figure 27. Training stage of the RL-TD3 agent for the correction of udref, uqref, and iqref.
Figure 27. Training stage of the RL-TD3 agent for the correction of udref, uqref, and iqref.
Energies 15 02208 g027
Figure 28. Time evolution for the numerical simulation of the PMSM control system based on the FOC-type strategy.
Figure 28. Time evolution for the numerical simulation of the PMSM control system based on the FOC-type strategy.
Energies 15 02208 g028
Figure 29. Time evolution for the numerical simulation of the PMSM control system based on the RL-TD3 agent for the correction of iqref.
Figure 29. Time evolution for the numerical simulation of the PMSM control system based on the RL-TD3 agent for the correction of iqref.
Energies 15 02208 g029
Figure 30. Time evolution for the numerical simulation of the PMSM control system based on the RL-TD3 agent for the correction of udref and uqref.
Figure 30. Time evolution for the numerical simulation of the PMSM control system based on the RL-TD3 agent for the correction of udref and uqref.
Energies 15 02208 g030
Figure 31. Time evolution for the numerical simulation of the PMSM control system based on the RL-TD3 agent for the correction of udref, uqref, and iqref.
Figure 31. Time evolution for the numerical simulation of the PMSM control system based on the RL-TD3 agent for the correction of udref, uqref, and iqref.
Energies 15 02208 g031
Figure 32. Time evolution for the numerical simulation of the PMSM control system based on control using SMC and synergetic controllers.
Figure 32. Time evolution for the numerical simulation of the PMSM control system based on control using SMC and synergetic controllers.
Energies 15 02208 g032
Figure 33. Time evolution for the numerical simulation of the PMSM control system based on control using SMC and synergetic controllers using an RL-TD3 agent for the correction of iqref.
Figure 33. Time evolution for the numerical simulation of the PMSM control system based on control using SMC and synergetic controllers using an RL-TD3 agent for the correction of iqref.
Energies 15 02208 g033
Figure 34. Time evolution for the numerical simulation of the PMSM control system based on control using SMC and synergetic controllers using an RL-TD3 agent for the correction of udref and uqref.
Figure 34. Time evolution for the numerical simulation of the PMSM control system based on control using SMC and synergetic controllers using an RL-TD3 agent for the correction of udref and uqref.
Energies 15 02208 g034
Figure 35. Time evolution for the numerical simulation of the PMSM control system based on control using SMC and synergetic controllers using an RL-TD3 agent for the correction of udref, uqref, and iqref.
Figure 35. Time evolution for the numerical simulation of the PMSM control system based on control using SMC and synergetic controllers using an RL-TD3 agent for the correction of udref, uqref, and iqref.
Energies 15 02208 g035
Figure 36. Time evolution for the numerical simulation of the PMSM control system based on control using SMC and synergetic controllers: TL = 2 Nm with 0.2 Nm magnitude of uniformly distributed noise and 50% increase in J parameter.
Figure 36. Time evolution for the numerical simulation of the PMSM control system based on control using SMC and synergetic controllers: TL = 2 Nm with 0.2 Nm magnitude of uniformly distributed noise and 50% increase in J parameter.
Energies 15 02208 g036
Figure 37. Time evolution for the numerical simulation of the PMSM control system based on control using SMC and synergetic controllers using an RL-TD3 agent for the correction of iqref: TL = 2 Nm with 0.2 Nm magnitude of uniformly distributed noise and 50% increase in J parameter.
Figure 37. Time evolution for the numerical simulation of the PMSM control system based on control using SMC and synergetic controllers using an RL-TD3 agent for the correction of iqref: TL = 2 Nm with 0.2 Nm magnitude of uniformly distributed noise and 50% increase in J parameter.
Energies 15 02208 g037
Figure 38. Time evolution for the numerical simulation of the PMSM control system based on control using SMC and synergetic controllers using an RL-TD3 agent for the correction of udref and uqref: TL = 2 Nm with 0.2 Nm magnitude of uniformly distributed noise and 50% increase in J parameter.
Figure 38. Time evolution for the numerical simulation of the PMSM control system based on control using SMC and synergetic controllers using an RL-TD3 agent for the correction of udref and uqref: TL = 2 Nm with 0.2 Nm magnitude of uniformly distributed noise and 50% increase in J parameter.
Energies 15 02208 g038
Figure 39. Time evolution for the numerical simulation of the PMSM control system based on control using SMC and synergetic controllers using an RL-TD3 agent for the correction of udref, uqref, and iqref: TL = 2 Nm with 0.2 Nm magnitude of uniformly distributed noise and 50% increase in J parameter.
Figure 39. Time evolution for the numerical simulation of the PMSM control system based on control using SMC and synergetic controllers using an RL-TD3 agent for the correction of udref, uqref, and iqref: TL = 2 Nm with 0.2 Nm magnitude of uniformly distributed noise and 50% increase in J parameter.
Energies 15 02208 g039
Figure 40. Comparison of the speed time evolution for the numerical simulation of the PMSM control system based on SMC and synergetic controllers using an RL-TD3 agent for the outer and inner loop correction.
Figure 40. Comparison of the speed time evolution for the numerical simulation of the PMSM control system based on SMC and synergetic controllers using an RL-TD3 agent for the outer and inner loop correction.
Energies 15 02208 g040
Table 1. PMSM nominal parameters.
Table 1. PMSM nominal parameters.
ParameterValueUnit
Stator resistance—Rs2.875Ω
Inductances on d-q axis—Ld, Lq0.0085H
Combined inertia of PMSM rotor and load—J8 × 103kg·m2
Combined viscous friction of PMSM rotor and load—B0.01N·m·s/rad
Flux induced by the permanent magnets of the PMSM rotor in the stator phases—λ00.175Wb
Pole pairs number—np4-
Table 2. PMSM control system based on SMC and synergetic controllers using RL-TD3 agent performances.
Table 2. PMSM control system based on SMC and synergetic controllers using RL-TD3 agent performances.
ControllersResponse Time [ms]Speed Ripple [rpm]
SMC and synergetic14.154.78
SMC and synergetic using RL-TD3 agent for the correction of iqref.13.754.15
SMC and synergetic using RL-TD3 agent for the correction of udref and uqref13.553.83
SMC and synergetic using RL-TD3 agent for the correction of udref, uqref, and iqref13.253.18
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Nicola, M.; Nicola, C.-I.; Selișteanu, D. Improvement of PMSM Sensorless Control Based on Synergetic and Sliding Mode Controllers Using a Reinforcement Learning Deep Deterministic Policy Gradient Agent. Energies 2022, 15, 2208. https://doi.org/10.3390/en15062208

AMA Style

Nicola M, Nicola C-I, Selișteanu D. Improvement of PMSM Sensorless Control Based on Synergetic and Sliding Mode Controllers Using a Reinforcement Learning Deep Deterministic Policy Gradient Agent. Energies. 2022; 15(6):2208. https://doi.org/10.3390/en15062208

Chicago/Turabian Style

Nicola, Marcel, Claudiu-Ionel Nicola, and Dan Selișteanu. 2022. "Improvement of PMSM Sensorless Control Based on Synergetic and Sliding Mode Controllers Using a Reinforcement Learning Deep Deterministic Policy Gradient Agent" Energies 15, no. 6: 2208. https://doi.org/10.3390/en15062208

APA Style

Nicola, M., Nicola, C. -I., & Selișteanu, D. (2022). Improvement of PMSM Sensorless Control Based on Synergetic and Sliding Mode Controllers Using a Reinforcement Learning Deep Deterministic Policy Gradient Agent. Energies, 15(6), 2208. https://doi.org/10.3390/en15062208

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop