1. Introduction
In recent years, machine learning and artificial intelligence have become highly discussed topics with significant advancements in these fields of computer science, automatic control, and robotics. Their applications are impressive [
1]. However, their deployment, particularly online, can still present challenges due to computational complexity. For example, training neural networks requires considerable time and resources, which may not always be feasible for real-time execution. In addition, many edge or executive devices, such as low-cost mobile robots, lack the necessary computational power. This is one of the reasons why these devices have become the focus of our research.
This paper addresses the problem of controlling a mobile robot with two degrees of freedom. The robot is a four-wheeled platform (two parallel axles) with a drive on the rear axle and steerable front wheels. Robot platforms of this type are very popular, and their models can be found in many references in the literature [
2,
3,
4]. Moreover, these models can be successfully applied for other vehicles with similar characteristics, e.g., bicycles. The control of mobile robots has gained considerable attention in recent years due to their wide range of applications in industries like manufacturing, logistics, autonomous vehicles, and service robots. One of the primary challenges in mobile robot control is ensuring precise trajectory tracking, particularly when robots are tasked with performing repetitive operations in uncertain environments where the design of feedforward signals is imperfect. To address the above challenges, several different strategies were proposed. Undoubtedly, model predictive control (MPC) constitutes an attractive solution due to its natural ability to incorporate various constraints. It can also be applied to changing environments. One possible strategy is developed in [
5]. It is based on a non-calibrated switching algorithm that determines the drive or brake adaptive regulation of MPC. It also avoids high-frequency oscillations by taking into account the braking torque. An alternative MPC solution is proposed in [
6]. It presents an MPC torque controller for velocity tracking, which compensates the nonlinearities caused by the forces of gravity, friction, and adhesion. Subsequently, the work [
7] tackles with an optimal velocity MPC under steering constraint, which is combined with the so-called pose correction. Finally, the paper [
8] deals with a compensatory lateral MPC technique for path tracking. It also combines MPC with an active disturbance rejection. In spite of the incontestable appeal of MPC, it involves the continuous solving of a related quadratic programming problem. Owing to the limited computational devices embedded within a mobile robot, it may be the limiting factor on the way toward the MPC application. Thus, an alternative solution is proposed, which is based on iterative learning control (ILC) [
9,
10,
11]. In particular, it can be applied to the mobile robots, which perform repetitive tasks (or trials) in such a way as to improve their trial-to-trial performance. Indeed, this technique is particularly useful for systems that perform the same task repeatedly, as it allows for progressive control refinement to minimize tracking errors. ILC was initially developed for robotic manipulator applications [
12] (iterative pick-up and placement task), but over time, the field has expanded significantly. Similarly to MPC, it can also deal with possibly time-varying constraints [
11]. Contrarily to MPC, it does not involve solving the quadratic programming problem. This feature recommends its possible application to mobile robots.
During each trial, information from previous trials is used to update the control signal, reducing the so-called tracking error (defined as the difference between the current trial output and the reference signal). Eventually, this process leads to the tracking error converging to an equilibrium point (zero in the case of linear deterministic systems). Two main approaches can be applied to ILC design: the two-stage approach [
10,
13] and the one-stage one. The two-stage approach assumes that the parameters of the ILC scheme are designed independently, while the one-stage design redefines the ILC scheme within the structure of linear repetitive processes (LRPs) [
14], applying stabilization results for that class of systems. For the one-stage ILC design, as mentioned earlier, the system is modeled as a closed-loop DLRP, where one of the state vectors represents the tracking error from the original system. It has been proven that using controllers that ensure stability throughout the trial process drives the tracking error to zero. The final challenge is how to compute the stabilizing controllers for the ILC scheme. To solve this problem, the linear matrix inequality (LMI) procedure [
15] can be utilized. As mentioned above, ILC is particularly appealing for mobile robots performing repetitive tasks, such as following the same path multiple times. It uses information from previous iterations (trials) to sequentially improve performance, making it ideal for applications such as automated delivery robots or warehouse navigation. However, vehicle control presents specific challenges due to the inherent nonlinearities in their dynamics. Traditional ILC designs often assume linear system dynamics or rely on simple linearization techniques. More advanced methods, known as the robust approach, address uncertainties and model disturbances [
16,
17,
18,
19]. However, these approaches may not fully capture the complexities of mobile robot behavior, especially when considering orientation-dependent kinematics. Recent advances have tackled this issue by developing linear parameter varying (LPV) models to better approximate nonlinear systems [
20].
The scheme of the proposed ILC strategy is depicted in
Figure 1. The focus is on developing an ILC scheme that incorporates a state-space model of the robot. Since the vehicle kinematics are nonlinear, particularly due to its orientation angle in the global reference frame, a tailored linearization strategy is applied. For each region, an LPV submodel is developed, allowing for more accurate approximations and a feasible controller design. Having such a model, it is possible to determine ILC gain parameters
and
(see
Figure 1), which make it possible to obtain a reference tracking ILC strategy. The paper is organized as follows.
Section 2 provides a necessary background about ILC. A mobile robot and its model are described in
Section 3.
Section 4 provides LPV-based ILC design.
Section 5 presents a set of case studies, which is followed by the experimental verification on a test platform (
Section 6). Subsequently,
Section 7 provides a suitable discussion about the results present in the paper. Finally, the last section concludes the paper.
2. Iterative Learning Control
As aforementioned in the ILC analysis, the finite-time iteration (or trial) dependence is used for control purposes. Let the subscript
k denote the trial number. This leads to the following state-space model of the considered system
where
,
is the time with the finite horizon
T and
,
,
are the state, the input and the output vectors, respectively.
denotes a derivative with respect to the time variable
Define also the reference signal
,
as a goal to be reached at the output and the consequence of the applied control strategy after a finite number of trials. To complete the model, it is necessary to provide the set of boundary conditions, i.e.,
where
are bounded functions. The control task can be alternatively defined as one allowing minimization of the tracking error, which is defined as
and we define the control signal in an iterative manner, i.e.,
where
is the input correction factor. Note that in (
4), it is assumed that the control signal used in the next trial is constructed as its counterpart in the current iteration improved by some signal generated basing on the control quality in the current trial.
Next, introduce the following signal
which defines the trial–to–trial state increment integral and is equivalent to
Combining (
1), (
4), (
5) and (
6) yields the following
Note, moreover, that
or
Hence, (
7) and (
9) can be gathered to provide the following state-space model
with the boundary conditions defined as follows
It is straightforward to see that the model (
10) takes the form of the specific subclass of 2D, i.e., DLRP [
11,
14]. There are
independent variables in the model – the time (which in this case is continuous) and the trial number (obviously discrete). The extended state vector contains two factors
and
, and the input signal is denoted as
. Let us assume that the control law obeys
Hence, (
10) can be rewritten in the following form
which is the closed-loop model of DLRP under consideration.
ILC Gains Design
As aforementioned, the ILC scheme (
13) takes the form of DLRP, and it is straightforward to see that the results developed originally for this subclass of dynamical systems can be applied to analysis and/or synthesis of the controller. Thus, since the tracking error
is part of the extended state vector, ensuring the stability along the trial (DLRP stability along the pass) guarantees that the error would tend to zero (according to stability theory). This in turn is equivalent to observing that the output
of the original system (
1) tends to the prescribed reference. In what follows, the problem of designing ILC gain matrices
and
is tackled. Their structure must ensure stability along the trial of the underlying ILC model. Hence, the following lemma can be recalled [
14].
Lemma 1. Stability along the trial of holds for (13) if, and only if,where and Note that Lemma 1 provides the necessary and sufficient stability condition. However, its practical applicability for the purpose of stability investigation and controller design is seriously limited. This is because the lemma is based, in fact, on performing an infinite number of numerical tests. Hence, the Lyapunov stability theory can be used in order to provide a more applicable condition. This leads to the task of solving a set of LMIs [
15]. However, it has to be underlined that stating the controller design problem in terms of LMI may appear with some conservativeness. Moreover, the well-known conditions are only the sufficient ones. Also, if the second block row of (
13) is considered, then it can be concluded that it represents a special case of a 1D discrete model. As a consequence,
tends to zero when this dynamical model is stable; that is, all eigenvalues of the block term
lie inside the unit circle of the complex plane. Note that the eigenvalues of
I are obviously equal to one; hence, the role of
is now to move the closed-loop matrix eigenvalues into the interior of the unit circle. Moreover, the distance of the largest module eigenvalue to the unit circle clearly plays an important role in the tracking error decay rate in the following trials. It is straightforward to see that the
based stabilization is possible if the pair
is controllable, i.e.,
and
denotes the length of the underlying vector. Note that this condition means that the full history of the current trial is available for the control purposes in the subsequent trial. However, when this condition is not true, and hence, the above rank condition is not satisfied, the ILC scheme should be redefined in such a way as to apply the feedback information from the more distant trials or more distant time moments (see, e.g., [
21]).
Let us recall the well-known results (see e.g., [
14]) regarding the control synthesis for DRLPs.
Lemma 2. Consider the ILC scheme of (13) given as a controlled DLRP. Then, there exist controller matrices such that the considered scheme converges to zero if there exist N and M of appropriate dimensions and the following LMI holdsIf this condition holds, then the control law matrices and are given by In (
17), Lyapunov decision matrices
Y and
Z are the same as those used for the controllers computation (
18), which can lead to an increase in the conservativeness of the LMI. To prevent this, so-called slack matrices can be introduced, as shown in Lemma [
22].
Lemma 3. Consider the ILC scheme of (13) given in terms of DLRP. Then, there exist controller matrices such that the underlying scheme converges to zero if there exist slack matrices , and N, M of appropriate dimensions, and the following LMI holds If this condition holds, then the control law matrices and are given by Finally, note that (according to [
22]) the values of gain matrices
and
obtained according to (
20) guarantee the asymptotic stability of the system (
13) fed by (
12). This simply guarantees that the tracking error
converges toward zero. This is possible due to the deterministic setting of the system (
1). Also, when (
19) is considered, it can be noted that
is responsible for the original system stability and
refers to the gradual decrease of the tracking error. Thus, to overcome disturbances and other unappealing effects, the proposed approach should be suitably extended. This issue will constitute a future research direction.
3. Mobile Robot Platform and Its Model
This study focuses on car-like mobile robots with Ackerman steering geometry, moving at relatively low speed (less than 0.5
). For this scenario, the forces acting on the wheels and therefore slip angles are expected to be negligible, which justifies the choice of a kinematic model. Like in [
2,
3], a four-wheel model was replaced with a bicycle model due to its simplicity (without the loss of generality). The considered plant is presented in
Figure 2. As it can be observed, Ackerman’s steering geometry forces all wheels to travel in circular paths around a common Instantaneous Center of Rotation (
), effectively preventing slippage. This configuration indicates that the steering angles of the right front wheel
and the left front wheel
are interdependent. Considering this relationship, they can be effectively represented by a single wheel in the bicycle model, which also travels around
and is controlled by the variable
. In order to calculate back the steering angles
and
on the basis of
, the following relationship can be used [
23]:
where
is the distance between the two front wheels. In practice, during the experiment described in
Section 6, a single steering value
is sent to the robot to control a servomotor, and then the mechanical link adjusts the angles
and
properly. This further justifies the use of the bicycle model.
Let us proceed to the kinematic model of the mobile robot considered. For that purpose, let us remind the lateral vehicle dynamics of the form [
24]
where
x,
y are the coordinates of the center of gravity (COG) of the vehicle,
is the vehicle heading angle (or orientation [rad],
v [m/s] is the linear velocity of the rear axle and
[rad/s] is the angular velocity with respect to the point
,
,
are the front and rear wheel steering angles, and
stands for the direction of the velocity at COG according to the longitudinal axis, which is simply called vehicle slip. For the purpose of further deliberations, the following assumptions are imposed:
- Assumption 1:
Slip angle is equal to zero, i.e., .
- Assumption 2:
Rear steering angle is equal to zero .
The first assumption is caused by the fact that an additional slip angle
estimator is required, which is not present in the current mobile robot setup. Note also that the robot will drive on a level surface with a constant rolling resistance parameter, and hence, such an assumption is fully acceptable. On the other hand, such a limitation can evidently impair the trajectory tracking, stability, and robustness to disturbances. Indeed, any surface slippery or disturbance may impair the robot performance. This topic is beyond the scope of the current research, and it will constitute the subject of future research directions. Indeed, one way to settle this issue is to treat disturbances and other unappealing effects like faults. As a result, the fault-tolerant control approaches can be applied [
25,
26]. The second assumption is due to the inability of the current setup to perform rear angle steering. As a consequence,
and
. Under the above stated assumption [
4], the nonlinear kinematic model is given as
while
is related to the steering angle
[rad] and the distance between the rear and front wheel axles
:
Note that velocity
v and the steering angle
are input signals of the bicycle model. One can see the nonlinear relationship between these variables to obtain
. Therefore, for the purpose of designing the controller,
will be taken as the second input of the system instead of
. In order to provide the control signal to the actual object, the value of
is recalculated based on the transformed equation (
24)
Finally, the position coordinates
of a robot and its orientation
can be defined as outputs.
Derivation of LPV Model
Note now that since (
23) is clearly nonlinear and does not comply with the (
1), in order to design the ILC scheme described in
Section 2, it is necessary to redefine the considered model. In what follows, it is clear that the input matrix can be treated as dependent on the parameter
. Hence, it is purposeful to present (
23) in terms of LPV model, i.e.,
Robot orientation is affected by the second input. Then, let us describe the parameter vector
. Arriving at the eventual system dynamics matrices,
Since all states are observable directly, matrix
C takes the identity form.
4. LPV Controller Design
Given the highly nonlinear, state-dependent nature of the state–space matrices, carefully selected controllers are to be found. A popular practice in robust control, where parameters of interest vary, is to find specific control gains that ensure coverage of the entire assumed range. Such designs have to be conservative, in other words, include extremum values that the plant may reach. An attempt was made to envelope (
27) with a simple polytope: square. This is equivalent to finding a controller for a set of fixed input matrices
that adhere to the linear process (
1), where all inequalities in (
17) must be jointly met. Finally, implementing such feasibility problems in a semidefinite programming solver, i.e.,
SeDuMi (
https://yalmip.github.io/solver/sedumi/, accessed on 20 December 2024) yielded mostly zero-filled (inapplicable) gains
,
i.e.,
This indicates conflicting relations of the system response to an input. The trigonometric functions in question change signs every time robot orientation
takes a value in a different quarter circle. A positive velocity increment contributes either positively or negatively toward position errors along axes X-Y. For this reason, four separate controllers that operate in disjoint circle sections are proposed. The angle of the sections is set to
. The selected nodes are shown in
Figure 3. As a result, gains for the first quadrant
were as follows:
To instantiate a different sector
of the circle,
Despite matrices being computed independently, their magnitudes are almost identical. Sign changes agree with the prior assumptions. A closer inspection of their values suggests that the feedback from signal
denoted by (
6) plays the role of an overshoot damper, counteracting input increments computed from error. Eventually, the derived controller was of the generalized DLRP form (
12)
with matrices being selected at every instant
t, which was ruled by division, as shown in
Figure 3.
8. Conclusions
This paper presents the design and implementation of the one-stage ILC scheme for a class of mobile robots. Since the original model of the robot is nonlinear, it is transformed into the equivalent LPV model containing the linear submodels. Then, for each submodel, the ILC design is defined, and an appropriate LMI is solved in order to provide the controller gain matrices. In the last step, they are applied in the control procedure. The proposed approach is attractive thanks to its application in the global reference frame, using measured states directly, without the need for local frame kinematic-based control. Moreover, experiments show the effectiveness of the proposed algorithm for the low-speed applications. It should be also noted that the preliminary results opened interesting research horizons: for instance, finer input nonlinearity division or expansion to velocity-dependant steering angle relation. A natural development would also be a comparison with the behavior of the full four-wheel model. A major milestone in future research will involve studying how skidding affects the learning process at higher driving speeds as well as strategies for preventing this phenomenon. This has implications for performance, wear, and safety. In addition, unexpected changes in the coefficient of friction of the surface (e.g., a wet surface) will be very important. Since these changes can occur in an unpredictable pattern between iterations (for instance, each time a mobile robot can skid on a different section of the path), future work should focus on extending the system model to include forces and incorporate robust control strategies with feedback from the current trial. The linearization of highly nonlinear systems has to be performed carefully, as the study has shown. Results from numerical procedures may not agree with requirements, and refinements might be necessary, such as the one we have developed and benchmarked. Finally, it should be noted that the current setup of the mobile robot is very simple and does not include such features like obstacle detection. This reduces the ability for including such knowledge in the trajectory-planning procedure. Thus, enhancing the mobile robot with such a feature can be one of the future research and development directions.