1. Introduction
Tikhonov regularization theory, introduced by Andrey Nikolayevich Tikhonov in the 1950s, provides a systematic introduction to the theory and applications of regularization [
1]. It addresses problems with unstable or no solutions, offering an effective mathematical framework to obtain stable and reliable solutions. The main idea is to introduce a regularization term into the cost function, balancing the trade-off between data fitting and solution stability. The regularization term, along with the data-fitting term, contributes to minimizing the cost function. The data-fitting term measures the difference between the solution and the observed data, while the regularization term constrains the solution’s characteristics, such as smoothness, sparsity, or minimum energy. By adjusting the regularization term, a balance can be achieved between data fitting and solution stability [
2,
3,
4,
5,
6].
The inverted pendulum is a highly typical nonlinear coupled system, and its study remains continuously relevant in research. Various fields, including intelligent systems, robotics, industry, and carbon reduction, can benefit from studying the inverted pendulum. This paper proposes the use of Tikhonov regularization theory combined with intelligent algorithms for the control of the inverted pendulum. The inverted pendulum is a classical nonlinear coupled system used to study oscillation and stability. It consists of a rod with a mass attached to one end, allowing it to rotate and swing along an arc path under the influence of gravity. The system’s dynamics are driven by gravity and the constraint force of the rod, resulting in various dynamic phenomena such as robotic motion, periodic motion, chaotic motion, and stability analysis. Due to its nonlinear nature, the inverted pendulum poses challenges for control and stabilization [
7,
8,
9,
10,
11]. Recent studies have explored the decoupling control of nonlinear systems using Tikhonov regularization theory. By introducing a regularization term, the decoupling of coupling relationships between system state variables is achieved, enabling the design of separate controllers for decoupling control [
12,
13,
14,
15]. Intelligent control systems for the inverted pendulum often rely on combinations of adaptive control and robust control. The coupling nature of the inverted pendulum system, where the motion of the rod influences the pendulum’s position and angle, makes controlling and stabilizing the system more challenging [
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27]. To address the challenges posed by the inverted pendulum system, this paper proposes the combination of a neural network with Tikhonov regularization theory for intelligent decoupling control. Sliding mode control (SMC) is employed for decoupling control in inverted pendulum systems to alleviate interactions among different variables. However, SMC encounters challenges, including computational burden [
18], slower convergence [
20], and singularity problems [
22]. This paper introduces aTikhonov-tuned sliding neural network (TSN) controller, which leverages sliding control for decoupling and incorporates Tikhonov parameters to train a neural network. These parameters are then integrated into the sliding surface to ensure system stability.
The problem statement of this paper revolves around achieving intelligent decoupling control for the inverted pendulum by combining a neural network with Tikhonov regularization theory. The primary objective is to develop an intelligent control approach that effectively decouples the system while maintaining stability through the sliding neural network which is adjusted by the Tikhonov regularization term. This research combines theoretical advancements with practical implementation to contribute to the field of intelligent control systems for the inverted pendulum, with potential applications in various industries and fields.
The contributions of this paper build upon the related literature on Tikhonov regularization, establishing its connections with sliding mode control and neural networks by using simulations and experiments. These foundations contribute significantly to the paper’s first contribution, as regularization in Tikhonov regularization and sliding mode control both aim to achieve robustness and stability in control systems. A critical aspect of this paper’s approach involves the selection of appropriate regularization terms, numerical simulations, and practical operations. This plays a crucial role in optimizing the performance of sliding mode control using neural networks.
2. Tikhonov Regularization Theory
In a standard linear regression model, the objective is to minimize the prediction mean squared error. However, if the data are limited or the number of features is large, the model can become overly complex and start fitting noise in the training data. This can lead to overfitting, where the model performs well on the training data but poorly on new, unseen data. To address the overfitting issue, we employed Tikhonov regularization. In this approach, we aimed not only to minimize the sum of squared errors but also to incorporate a regularization term into the loss function. This regularization term penalized large coefficient values [
1,
2,
3,
4,
5,
6].
For example, let us consider a linear regression model with two features
, where the model equation is as follows:
where
and
are the coefficients of the features,
b is the intercept term, and
is the model output.
In Tikhonov regularization, the loss function will include an additional term proportional to the L2 norm of the coefficients:
where
is the number of data points,
is the predicted value,
is the actual value, and
is the regularization term. The purpose of this Tikhonov equation is to balance the accuracy of the predicted values with the complexity of the model. The first term, the mean squared difference, measures the discrepancy between the predicted and actual values, emphasizing the importance of accurate predictions. The second term, the regularization term, helps to control the model’s complexity by penalizing large weight values. The regularization term
determines the trade-off between fitting the data well and keeping the weights small. A higher
value increases the regularization effect, promoting smaller weights and potentially reducing overfitting. Conversely, a lower
value gives more weight to accurate predictions, allowing for larger weight values. By optimizing this Tikhonov loss function, the goal was to find the best combination of weight values that minimized the overall loss and achieved a balance between accurate predictions and model complexity.
By including the regularization term, the model was encouraged to keep the coefficients ( and ) small while still fitting the data. The values for and in the Tikhonov equation were typically obtained through an optimization process. The specific approach to obtaining these values depended on a gradient descent approach, which is an iterative optimization algorithm that aims to find the minimum of a function. In the context of the Tikhonov equation, gradient descent can be used to update the values of and by iteratively adjusting them in the direction of the steepest descent of the loss function. The regularization term controls the degree of regularization applied. A larger value leads to smaller coefficient values and simpler models. In summary, Tikhonov regularization penalizes large coefficient values in a linear regression model, helping prevent overfitting and making the model more balanced, with better generalization ability.
To illustrate this concept, let us demonstrate the prediction of a sine function using the Tikhonov regularization method. First, we generated data for the sine function and defined a Tikhonov regularization function. This function utilized matrix operations to compute the regression parameters, denoted as w. Subsequently, we set the regularization term
and constructed a design matrix X of size n. Utilizing Equation (2), we calculated the minimum Loss to use the Tikhonov regularization function for prediction and stored the results in y_pred. Finally, we plotted the original sine function and the predicted results of Tikhonov regularization as the blue and red curves, respectively (refer to
Figure 1). Tikhonov regularization has been proven to be a valuable technique in control systems, offering stability and reliable solutions for problems with unstable or nonexistent solutions.
When performing polynomial fitting, as the polynomial degree n was set to 5 or higher, the polynomial fitting model became very flexible and could fit every point in the training data, including the noise. This led to the black sine function curve on the graph almost overlapping with the “ = 0.01” red dash line, which is overfitting. However, such a model may have poor predictive capability on new data because it focuses too much on specific features of the training data and fails to capture the overall trend.
To address the overfitting issue, we could adjust the regularization term
to a value of 10, as shown by the purple dashed-double-dot line at the bottom. A larger value of
penalized the model’s parameters more strongly, reducing the risk of overfitting. We could try adjusting the value of
to find the optimal balance point, such as
= 0.1, as shown by the blue dot line in the middle, which could achieve a balance between fitting the training data and the predictive performance on new data.
Figure 1 shows the simulation results of adjusting the value of
to find the optimal equilibrium point, avoiding overfitting and achieving the best prediction of the sine wave. The different
yielded a different system performance, as demonstrated in
Figure 1. By varying the
value, we could observe different levels of system stability, response time, overshoot, and other performance metrics. This allowed for fine-tuning of the system’s behavior based on the specific requirements of the application. Therefore,
Figure 1 serves as a visual representation of how different
values impacted the system’s performance, providing valuable insights for selecting the most suitable
value in practical applications.
3. Intelligent Algorithm Design
The inverted pendulum represents a classic control system problem, wherein a pole is mounted onto a moving cart, and the objective is to keep the pole upright. The primary task of the controller is to generate control signals based on the system’s state, including the pole’s angle and the cart’s position, to maintain balance.
Typically, we would establish a mathematical model to describe the dynamics of the inverted pendulum and use that model for controller design. Tikhonov regularization can be applied to the parameter estimation problem in controller design. In this case, the cost function could be formulated to minimize the error between the system output and the desired output, while incorporating a Tikhonov regularization term to constrain the size and complexity of the parameters. By adjusting the regularization term, a balance could be achieved between fitting the observed data and ensuring the smoothness of the control parameters. Therefore, the first step involved simplifying the nonlinear coupled system into a linear system for design purposes, while the remaining nonlinear characteristics were addressed by the sliding neural network controller. For the linear parts, LQR (linear quadratic regulator) control could be used. For the nonlinear components, Tikhonov regularization could be applied to tune the weights of the neural network. For sliding mode control, the sliding surfaces contributed to the improved performance and robustness of the inverted pendulum system. This enabled adaptation to uncertainties, noise, and model errors through parameter estimation and facilitated the generation of suitable control strategies to stabilize the inverted pendulum and enhance overall performance.
Therefore, this paper proposes the design method of the Tikhonov sliding neural network (TSN) controller, as shown in
Figure 2, which combines Tikhonov regularization with a neural network to achieve intelligent control of the inverted pendulum. In this section, a step-by-step explanation is provided of how to combine and compute Tikhonov regularization, neural networks, and sliding mode control. The process starts by setting the goal of achieving intelligent control for nonlinear systems through the application of Tikhonov regularization within the framework of sliding mode control. A sliding mode controller is then established to control the system by sliding its state onto a specific sliding surface to achieve the desired control effect. Neural networks are introduced as a crucial component to monitor and adjust the regularization terms in Tikhonov regularization, utilizing their learning and adaptive adjustment capabilities. Finally, the combination of Tikhonov regularization, neural network, and sliding mode control is realized through iterative computation and application of the controller, which involves integrating the system’s inputs and outputs with the controller and adjusting the regularization terms and neural network weights based on the specific circumstances.
3.1. Linear Model
This subsection indeed serves as a foundation for introducing our proposed nonlinear control method. In this subsection, we establish the groundwork by presenting classical optimal control techniques.
Consider a mathematical linear model presented as follows:
where
and
are system dynamics matrices, while
and
are all m-dimensional vectors of time
. To introduce the controller into the system, we can represent the control input as
, where
is the sought controller gain matrix. Substituting u back into the dynamics model, we obtain the following:
Below, we will briefly introduce one method to derive the solution for
. Suppose our goal is to design a state feedback controller that keeps the angle of the inverted pendulum near the target angle. According to the LQR theory [
8,
25], our control objective can be defined as minimizing the cost function
:
where
and
represent the state weight matrix and control weight matrix, respectively, both having a square dimension of m. We want to adjust the controller gain matrix
to minimize
. The optimal solution can be derived through algebraic or differential methods in optimal control theory. One common approach is to solve the algebraic Riccati equation [
26]. By solving the following Riccati equation, we can obtain the optimal controller gain matrix
:
where
is the symmetric positive definite state feedback gain matrix with a square dimension of ‘m’. Once it is determined, the controller gain matrix
can be selected using the following formula, ensuring that the system described by Equation (6) has negative eigenvalues to maintain the stability of Equation (4). The detailed derivation process is presented in
Appendix A.
Thus, the optimal state feedback controller can be expressed as:
3.2. Nonlinear Model
From Equation (5), this approach optimizes the controller’s performance by minimizing the cost function J and guiding the system’s state towards the desired stable point. However, in certain cases, it may be necessary to introduce a regularization term into the controller design to address specific issues. For example, when the system’s model contains uncertainties or noise, a sliding mode neural network can be employed to estimate the controller’s parameters, thereby improving robustness and performance.
A canonical nonlinear system is represented as below:
where
,
,
, and
are the state vector, dynamic with respect to
; control input vector; and disturbance vector, respectively. Each of these vectors has a dimension of ‘m’ and varies over time.
is an m-dimensional square system matrix. The unknown disturbance is assumed to have a known limited norm bound, i.e.,
. The control law can be deduced from Equation (9), and the expression is as follows:
Then, according to the universal approximation theorem [
25], there should be an optimal control law, which is expressed as follows:
by defining the tracking error vector and the error dynamic as below:
where
.
If we assume that is equal to (as indicated in Equation (13)), we can then select an appropriate value of as defined in Equation (8) to ensure the convergence of .
However, in practice, the
and
are always unknown, making the design
of Equation (12) challenging. Many recent studies have attempted to address this problem, but they encounter issues with complex manipulations and inadequate performance. In this paper, we propose a simple approach to designing the controller. We utilize a Tikhonov-tuned sliding mode neural network (TSN) controller to handle model-free problems, ensuring that the states closely approach the desired region. Simultaneously, a sliding mode controller is employed to maintain the system’s states on the sliding surface, eventually allowing
to approximate
, leading to improved performance and making sure that the states are converged. The designed controller is as follows:
where
represents an LQR linear controller discussed in
Section 3.1, while
denotes a TSN controller. The proposed TSN control system is introduced in
Figure 2 and
Figure 3 and will be further elaborated in the following sections.
3.3. Tikhonov-Tuned Neural Network (TNN) Architecture
Figure 3 depicts a four-layer Tikhonov-tuned neural network (TNN). It comprises the input, activation, rule, and output layers. The energy function is minimized using the gradient descent method to obtain optimal weights for the TNN. Signal propagation and basic functions in each layer of the TNN are introduced as follows:
(a) Input Layer: In this layer, each node is associated with net input and net output values, denoted as
and
, respectively.
where
represents the
i-th input to the node of Layer 1 and m is the number of input variables. The link weights in this layer are all set to unity.
(b) Activation Layer: Each node in this layer performs an activation using the ReLU (rectified linear unit) function, which is defined as
[
28]. The derivative of ReLU is 1 when the input number is greater than zero; otherwise, it is equal to 0. ReLU will be used as the activation function for the
j-th node of the
i-th input.
(c) Rule Layer: The cardinality of this layer determines the count of fuzzy rules, and its neuron count profoundly influences the rule complexity. Each individual node, represented by
, plays a pivotal role as it multiplies the incoming signal, giving birth to the emergent output specific to the
j-th rule node. This dynamic interaction fosters diverse rule behaviors, leading to intricate and versatile system responses. The formal expression is given by the following equation:
(e) Output Layer: This comprises multiple nodes, each referred to as an output node, responsible for performing the crucial defuzzification operation. Among these nodes, we have a particular one labeled as
which calculates the overall output by summing all incoming signals. This summation process is denoted by the following equations:
In Equations (19) and (20), we can further describe the components involved in this calculation:
: The link weight, representing the output action strength of the j-th column output associated with the k-th rule.
: The j-th input to the node of Layer 4.
: The total number of output nodes.
: The k-th column output of the TNN controller.
ReLU activation has the property of avoiding the vanishing gradient problem and promoting faster training by allowing gradients to flow when inputs are positive. As a result, the TNN architecture can handle zero input values and propagate gradients effectively, even in the absence of bias neurons.
Furthermore, Tikhonov regularization factors are incorporated into the TNN to ensure stability. These regularization factors help prevent overfitting and improve the generalization capability of the network.
3.4. Neural Network Controller Design
According to the gradient theory, we consider a single neuron to derive the equations in the neural network. To tune the weights of the TNN, the energy function and cost function of the Tikhonov are defined as follows to employ the gradient operation:
where
is the Tikhonov regularization term and
is the cost function of the Tikhonov function. Thus, the total energy function can be presented as follows:
The gradient calculus of
with respect to
can be obtained as follows:
where
is the learning rate. Similarly, the activation layer update law can be as follows:
Up until this point, the derivations are complete, and the entire TNN design is now finalized. The weight update laws for the activation layer and rule layer are presented in Equations (24) and (25), respectively. These two layers play a crucial role in the TNN. Additionally, the Tikhonov regularization term is incorporated to prevent overfitting. Moreover, the model demonstrates the capability to predict the next state of the system and generate appropriate control forces to address uncertainties and disturbances.
In the derivation process, the variable is used. However, in the subsequent practical application, the error variable is substituted for in calculations.
3.5. Sliding Mode Controller Design
From Equation (12), according to the universal approximation theorem [
25], suppose there exists an optimal
such that
where the term
denotes the approximation error.
is the LQR controller discussed in the previous section.
By the substitution of (26) and (14) into (13), the error dynamics become
where
is assumed to be bounded by
, where
is a carefully selected positive constant vector. The sliding mode controller aims to mitigate the influence of the approximation error. In the forthcoming section’s demonstration, the TSN controller can be achieved in the following manner:
where ‘
’ represents the sign function, which can be replaced with a saturation function to mitigate chattering effects. Furthermore,
is a symmetric positive definite matrix.
3.6. Stability Analysis
According to the stability theorem of nonlinear systems [
25], let us define the Lyapunov function as follows:
The derivative of the Lyapunov function can be expressed as follows:
From Equations (27) and (30), we arrive at the following equation:
where
is a positive definite symmetric matrix and
is a Hurwitz matrix. Matrix
can be utilized to represent the common part of
and
in Equation (31), which is given by
Substituting Equation (28) into (31) will result in the following:
where
is the minimum eigenvalue of the matrix
. We can select a large eigen diagonal matrix and the positive vector
in the context of Lyapunov stability. The choice of a large eigen diagonal matrix is indeed a key factor in ensuring the Lyapunov stability of the system. In summary, the choice of a large eigen diagonal matrix is a deliberate strategy to ensure the Lyapunov stability of the system, and the selection of the positive vector
is a critical part of this analysis, but it may vary depending on the context and specific stability requirements of the problem at hand. Since the error vector
is measured using the 2-norm, the right-hand side of Equation (27) remains bounded, ensuring a stable behavior for our system. Additionally, based on Equations (29) and (33), we can find the Lyapunov function
using the 2-norm and
infinity norm, respectively. Applying Barbalat’s Lemma, which guarantees the convergence of states, we confidently conclude that the error vector
will approach zero, demonstrating the stability of the system.
The aforementioned design methods are summarized in
Figure 4, with each block marked with its corresponding equations.
4. Simulation Result
4.1. Example 1: Cart Inverted Pendulum
By analyzing the linear quadratic regulator (LQR) theory, we could determine the stability of the linear model and obtain the controller K. Consistently observing a negative derivative in Lyapunov analysis indicates stability in the nonlinear system. Introducing Tikhonov regularization allowed us to fine-tune the weights of the TNN in order to prevent overfitting and achieve accurate predictions to compensate for the control forces of the nonlinear system, especially in cases of coupled systems. The training objective was to drive the sliding surface, influenced by various errors, to converge towards zero.
Consider the model of the inverted pendulum as below: [
25]
where
,
is the angle from the vertical, and
is the control force (Newton). The simulation model for an inverted pendulum was utilized as described below and adhered to the specifications outlined in
Table 1.
Based on the procedure outlined in
Section 3.1, the linearized model around the original point couldbe obtained as follows:
Subsequently, the values of
couldbe determined as
= [−258.12 −170.35]; then, the eigenvalue of
was [−1.0632 −34.7103]. As shown in
Figure 2, the control system employed a sliding neural network to handle the nonlinear coupling characteristics and dynamically adjust the weights of the TNN to ensure both stability and predictability. After comparing the LQR, sliding neuralnetwork (SN), and Tikhonovsliding neural network (TSN) methods, as depicted in
Figure 3, under the condition of requiring both the position and the angle to be zero, it becameevident that the TSN method excelled in decoupling. The weight factor values, presented in matrix format, were derived from the aforementioned sections. These values were stored in a matrix, where each row represented a distinct factor or variable, and each column corresponded to the relevant weight factor value. The optimal regularization term
was set to 0.1, and initial weightings were all set to 0.15. According to Equation (28), the matrics of
,
, and
.
Performance indices serve as quantitative measures to evaluate the performance of the simulation of an inverted pendulum system. These indices provide insights into various aspects, including stability, tracking accuracy, control effort, settling time, and energy efficiency.
Figure 5 clearly demonstrates that the TSN controller outperformed both the LQR and theSN controllers.
4.2. Example 2: Rotary Inverted Pendulum
The rotary inverted pendulum represents a unique variant of the inverted pendulum, distinct from the cart inverted pendulum, which is governed by the linear back-and-forth motion of a cart. To assemble the rotary inverted pendulum, we affixed a connecting pivot arm to a servo motor’s shaft, employing rotational torque to manage another connecting pendulum fastened to its trailing end. When the connecting pendulum remained motionless, it hung vertically due to gravity. Upon applying a driving torque, it overcame gravity, assuming an upright vertical position. Consequently, it did not follow a linear path but instead executed a circular rotation. The control principles for this system resemble the control laws detailed in this paper. Both systems are interconnected, nonlinear systems, where a single input governs two outputs. The connecting pivot arm’s position transitioned from 0 degrees and eventually returned to 0 degrees. The pendulum’s position oscillated from 0 degrees to either positive or negative 180 degrees, contingent on its initial rotation direction, whether clockwise or counterclockwise.
Figure 6 illustrates the architectural diagram.
The equation representing the rotary inverted pendulum model is provided below [
29]:
where
,
,
,
,
,
,
,
,
, and
. For a comprehensive list of simulation and physical parameters, please refer to
Table 2.
As shown in Example 1, following the procedure outlined in
Section 3.1, we couldobtain the linearized model around the original point in the following manner:
Utilizing the identical approach as illustrated in Example 1, we could obtain the feedback gain as
= [−9.8 −114.4 −6.22 −15.48]. The optimal regularization parameter was configured at
=0.05, and initial weightings were all set as 0.1.According to Equation (28), the matrics of
,
, and
. The simulation began with initial conditions set to
and
, and the results are compared to those of LQR and SN in
Figure 7.
4.3. Example 3: Arm-Driven Inverted Pendulum
The arm-driven inverted pendulum is a robotic system consisting of two freely rotating links. These links operate on a common plane, with the lower motor applying torque to rotate the arm, which, in turn, drives the upper link. The primary objective of this system is to maintain equilibrium, ensuring both poles remain upright and balanced without toppling over. Due to the influence of gravity on both connecting rods and the fact that only one force propels them, nonlinear characteristics interplay and mutually affect each other. Consequently, controlling this system proves to be more challenging compared to a rotary inverted pendulum, as it represents a highly coupled nonlinear system. It necessitates not only preventing the driving link from falling but also ensuring the arm remains stable. In this paper, we propose a method to address the issue of coupling in this system. The architecture diagram of arm-driven inverted pendulum is shown in
Figure 8.
where
,
,
,
,
, and
. All parameters were identical to those in example 2, specifically in
Table 2. Following the same procedure as in example 2, we linearized the model around the original point as shown in Equation (40).
The feedback gain was determined as follows:
= [69.16, 141.81, 1.44, 9.78]. We set the optimal regularization parameter to
= 0.03, and initial weightings were all set as 0.1. According to Equation (28), the matrics of
,
, and
. The simulation was initiated with initial conditions
,
,
, and
, which were defined as in
Figure 8 and set to
and
; the results were compared to those obtained using the LQR and SN methods, as shown in
Figure 9.
The simulation results reveal that both the cart and rotary inverted pendulums exhibit low coupling, which makes it difficult to distinctly demonstrate the advantages of the TSN control law when compared to the other two control laws. However, in the case of the arm-driven inverted pendulum system, which is a highly coupled nonlinear system, the superiority of TSN becomes evident.
During the arm-driven inverted pendulum simulation, we observed that the LQR control exhibited oscillations before converging to the steady state. In contrast, the SN control law displayed a longer settling time and higher overshot in transient response, making it less effective than the TSN control law. The key differentiator lies in TSN’s utilization of Tikhonov’s regularization term, which enhanced the decoupling performance by adjusting the weight of the neural network to improve prediction.
In the following section, we will conduct practical verifications of two types of inverted pendulums: rotary and arm-driven. Subsequently, we will demonstrate the robustness and decoupling performance of TSN, compared to implementing the other two methods, LQR and SN. The LQR and SN controllers face difficulties in real empirical experiments due to factors such as vibration and the precision of encoders. These factors can indeed render the system unstable in the steady state.
6. Conclusions
This paper proposed a method that combines the Tikhonov regularization algorithm with sliding neural network control for nonlinear coupled systems. The purpose was to explore methods for stabilizing an inverted pendulum, enabling it to exhibit AI (artificial intelligence)-like stability and predictive capabilities in articulated actuators. Through mathematical analysis, simulations, and practical implementations, the results demonstrate compliance with stability and performance requirements, thus proving the feasibility and superiority of this approach.
The Tikhonov regularization, sliding surface, and neural network methodology proposed in this work differs from recent research in several aspects. Firstly, it provides a comprehensive and integrated approach by combining multiple control methodologies, leading to improved system performance and stability. Secondly, the use of the Tikhonov regularization term helps in effectively adjusting the weights of neural networks and allows for adaptive learning and real-time adjustments to enhance decoupling ability. Thirdly, the incorporation of sliding mode control achieves desired control objectives and enhances the controller’s capability of handling uncertainties and disturbances. Overall, this proposed methodology offers a novel and effective solution for decoupling control in the inverted pendulum system, showcasing its superiority and reliable robustness in decoupling.