Next Article in Journal
An Improved Dung Beetle Optimizer for the Twin Stacker Cranes’ Scheduling Problem
Previous Article in Journal
Lunarminer Framework for Nature-Inspired Swarm Robotics in Lunar Water Ice Extraction
Previous Article in Special Issue
A Motion Planner Based on Mask-D3QN of Quadruped Robot Motion for Steam Generator
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bipedal Stepping Controller Design Considering Model Uncertainty: A Data-Driven Perspective

State Key Laboratory of Robotics and System, Harbin Institute of Technology, Harbin 150080, China
*
Authors to whom correspondence should be addressed.
Biomimetics 2024, 9(11), 681; https://doi.org/10.3390/biomimetics9110681
Submission received: 5 September 2024 / Revised: 27 October 2024 / Accepted: 1 November 2024 / Published: 7 November 2024
(This article belongs to the Special Issue Bio-Inspired Locomotion and Manipulation of Legged Robot: 2nd Edition)

Abstract

:
This article introduces a novel perspective on designing a stepping controller for bipedal robots. Typically, designing a state-feedback controller to stabilize a bipedal robot to a periodic orbit of step-to-step (S2S) dynamics based on a reduced-order model (ROM) can achieve stable walking. However, the model discrepancies between the ROM and the full-order dynamic system are often ignored. We introduce the latest results from behavioral systems theory by directly constructing a robust stepping controller using input-state data collected during flat-ground walking with a nominal controller in the simulation. The model uncertainty discrepancies are equivalently represented as bounded noise and over-approximated by bounded energy ellipsoids. We conducted extensive walking experiments in a simulation on a 22-degrees-of-freedom small humanoid robot, verifying that it demonstrates superior robustness in handling uncertain loads, various sloped terrains, and push recovery compared to the nominal S2S controller.

1. Introduction

Reduced-order models have always played a crucial role in generating gait for humanoid robots. They capture the key physical characteristics of walking in humanoid robots, and due to their low-dimensional state and control variables, they can typically be computed in real time. Linear inverted pendulum (LIP) models are commonly used to generate feasible foot placements and have been widely studied in the robotics community over the past few decades [1,2]. The restrictive assumptions of LIP, such as fixed height and the neglect of angular momentum, lead to varied degrees of deviation in robots with different mass distributions. A notable issue is how to incorporate this model uncertainty into the design of the stepping controller and improve its robustness for walking.
The early LIP models are closer to quasi-static locomotion, requiring the center of pressure to remain within the convex hull formed by the foot contacts. More dynamic locomotion benefits from the introduce of the capturability theory and the viability theory, leading to the development of footstep planning based on the divergent component of motion (DCM) [3]. The idea of the DCM-based controller is to introduce constraints for the unstable components of LIP dynamics, thereby stabilizing the overall dynamics. The problem with this approach is that the LIP model and the constraints formulated for it both deviate from the original system. A possible direction is to use less conservative robust model predictive control (MPC) [4] by considering the effect of future deviation accumulation through the concept of constraint tightening. More precise reduced-order models (ROMs) include centroidal dynamics [5], and its simplified version, the single-rigid-body dynamics model, has shown impressive performance in bipedal robots [6], but they focus on solving for reference contact wrench or centroidal momentum. Explicitly incorporating contact point planning into the solution introduces the complementarity relaxation problem, which is still challenging to solve in real time [7]. In some works, the constraints of the LIPM have been relaxed, for example, the introduction of variable-height capturability analysis in [8], which, however, requires a dedicated sequential quadratic programming solver for nonlinear optimization problems. The improved model proposed in [9] extends the LIPM to adapt to low-friction environments and designs a stepping controller for a planar five-link point-foot robot. Additionally, in recent years, the fusion of multi-strategy approaches (stepping, ankle, hip) has become a research hotspot [10,11]. However, these methods typically utilize multiple reduced-order models to represent different inputs (such as ZMP and equivalent ground reaction forces). How to address the dynamic and kinematic consistency between different models still requires further research.
The step-to-step dynamics introduced in [12] provides another perspective for stepping controllers. Its concept aligns with the Poincaré return map [13] in the control community, indicating that the stability of periodic walking is not instantaneous, but is stable from footstrike to footstrike. The Poincaré map maps the state x k at a specific point on the Poincaré section within one step to the state x k + 1 at the same point in the next step. The periodic orbit creates a fixed point, and the stability of the orbit is equivalent to the stability of the Poincaré map.
The S2S dynamics of the full-order model are high-dimensional and nonlinear. Therefore, Ref. [12] proposed using HLIP for approximation, ensuring the tracking error converges to an error invariant set to achieve orbital stability. To improve the approximation accuracy of the model, linear regression [14] and support vector machines [15] are used to learn a new linear or polynomial S2S dynamics. These methods have improved the tracking accuracy of the controller to some extent; however, when approximating with a single time-invariant linear system, errors always exist, especially when walking speed increases, height changes, and load varies. The question of whether it is possible to design control law that fully consider such uncertainty and provide strict robustness guarantees is the focus of this paper.
Direct data-driven control based on behavioral system theory [16] provides a new possible perspective for this problem. Unlike traditional adaptive control [17], direct data-driven control bypasses the identification step, constructing feedback control laws directly from data while guaranteeing the stability and performance of the unknown system. The fundamental results of Willems et al. [18] on noiseless data of linear systems have led to widespread research in the control community in recent years [19]. Designing robust feedback controllers using noisy input-state data is a current research hotspot in this field. Linear matrix inequalities (LMIs) are used to provide robust guarantees for data-consistent uncertain linear systems, such as the S-lemma [20] and Petersen’s lemma [21]. Improving the dynamic performance of control systems is also being studied, such as adjusting the LMI region [22] and enhancing H 2 [23] or H performance [20].
Contribution: In this letter, we directly construct a robust state-feedback controller using input-state data of BRUCE’s flat-ground walking collected from simulations with a nominal controller. The deviation between the linear system and the nonlinear system due to model uncertainty can be considered as uncertain noise, which is approximated by a bounded energy ellipsoid. By using robust data-driven control (RDDC) [21], we design a semi-definite programming (SDP) problem to solve for the state-feedback gain that stabilizes a set of data-consistent ( A , B ) . Here, ( A , B ) represents a family of matrices equivalently characterized by bounded energy ellipsoids, rather than the deterministic ( A , B ) obtained through traditional parameter identification or the ( A , B ) used in HLIP. We provide a detailed explanation in the following text for Equation (5). We integrate the data-driven stepping controller into the whole-body control (WBC) to generate joint torques, and verify the improvement in robustness performance through experiments in different scenarios, such as push recovery, varying loads, and slopes, just as is shown in Figure 1.

2. Bipedal Robot Model and Control Framework

2.1. Bipedal Robot Model

BRUCE [24] is a small, low-cost humanoid robot consisting of a torso and two 5-DoF legs. Presented in Figure 2, each leg includes a spherical hip joint, a single-DoF knee joint, and a single-DoF ankle joint. Each foot is designed to make line contact with the ground to prevent overconstraint. Each joint is equipped with proprioceptive drive using BEAR actuators [24]. Relevant physical parameter details are listed in Table 1.

2.2. Control Framework Overview

BRUCE’s actuator is positioned near the hips, thus reducing the leg inertia. However, as shown in Table 1, the inertia of the legs relative to the torso is still significant, which actually contradicts the assumptions of LIP. This introduces model uncertainties, motivating us to introduce robust data-driven control.
The entire control framework is illustrated in Figure 3. We can use a nominal controller to make BRUCE walking in simulation and collect the dataset D origin of CoM state and step-length pairs. Gaussian noise with bounded amplitude is added to D origin to obtain the noisy dataset D noise . Offline, we solve for robust state feedback gains K. During online operations, the Stepping Controller is used to solve for step length in real time, and a fifth-order polynomial interpolation is employed to obtain foot trajectories. At the low level, the Whole-Body Controller calculates joint torques to track the trajectories in task space.

3. Robust Stepping Controller Design

3.1. S2S Dynamics in Data-Driven Perspective

The dynamics of bipedal walking are hybrid, encompassing both continuous dynamics from swing leg lift-off to near landing (pre-impact) and discrete switching dynamics around ground contact. For ROMs like LIP, we consider only the evolution of the center of mass (CoM), where a complete S2S dynamics describes the state transition from the k-th step’s pre-impact state to the ( k + 1 ) -th step’s pre-impact state. The dynamics in the x and y directions are decoupled and can be approximated separately as
x k + 1 = A x k + B u k
where x k R 2 represents the position of the CoM relative to the stance leg contact point and the CoM velocity (or alternatively, the momentum of the CoM relative to the contact point) in the decoupled x/y-axis. u k R 2 represents the step length.
Remark 1.
A R 2 × 2 and B R 2 × 1 represent arbitrary linear approximations of S2S dynamics. For the LIPM/ALIP forms, please refer to [12] and [25], respectively.
In this letter, we provide the deterministic data-driven form of A * and B * . Assume the pre-impact state and step length data collected at each step from simulations form the following matrices:
X 0 : = [ x d ( t 0 ) , x d ( t 1 ) x d ( t T 1 ) ] U 0 : = [ u d ( t 1 ) , u d ( t 2 ) u d ( t T 1 ) ] X 1 : = [ x d ( t 1 ) , x d ( t 2 ) x d ( t T 1 ) ]
The error of the data-driven S2S dynamics relative to the real system’s S2S dynamics can be regarded as an unknown disturbance sequence, denoted by W 0 : = [ ω t 0 , ω t 1 , ω T 1 ] . W 0 satisfies
X 1 = A * X 0 + B * U 0 + W 0
Assume the disturbance sequence W 0 has bounded energy, i.e., W 0 W , where, for some matrix Δ ,
W : = W R n × T : W W T Δ Δ T
The set C of all matrix pairs consistent with the data (2) and set W can be represented as
C : = [ A B ] : X 1 = A X 0 + B U 0 + W , W W
We note that [ A * B * ] C , which is a matrix ellipsoid. At this point, we have obtained a viewpoint that differs from the conventional notion of S2S approximation. Specifically, S2S dynamics are represented by a set of matrices C , rather than a specific A and B. From the control synthesis perspective, we also need to stabilize the entire family of systems contained in this set.

3.2. Target Pre-Impact States in Data-Driven Form

As shown in Equation (1), the input to the S2S dynamics of the bipedal robot is only the step length. Therefore, it is necessary to adjust the step length to achieve the desired CoM state. From the perspective of periodic stability [12], a stable periodic walking pattern is achieved by solving for the step length to ensure that the robot returns to the current pre-impact state after completing 1 step (or 2 steps) with x k = x k + 1 ( x k = x k + 2 ), thus generating the P1-Orbit and P2-Orbit [12]. For an intuitive explanation and the physical significance of the two orbits, see Figure 2. In the design of the stepping controller in this paper, the P1-Orbit is used to generate forward and backward walking, while the P2-Orbit is used to generate lateral walking (for lateral walking, it takes two steps to return to the initial state). Refer to [12] for a detailed explanation of this setting.
Lemma 1
([21]). According to the certainty equivalence principle, the matrix ellipsoid center of (5) is actually equivalent to the solution of the following least squares problem:
[ A l s , B l s ] : = arg min [ A B ] | | X 1 A X 0 B U 0 | | F 2 = X 1 X 0 U 0 = [ A * , B * ] + W 0 X 0 U 0
where represents the Moore–Penrose pseudoinverse of the matrix. According to the definitions of the P1-orbit and P2-Orbit, we can obtain the target CoM states of the two orbits given the desired step length u * :
x p 1 * = A l s x p 1 * + B l s u *
x p 2 * = A l s ( A l s x p 2 * + B l s u l / r * ) + B l s u r / l *
For the desired velocity usr-input v d and swing time T s s p , in the case of the P1-Orbit, u * = v d · T s s p , and for the P2-Orbit, u l / r * + u r / l * = v d · T s s p . We can solve the above equations to obtain the target CoM states for the two orbits, respectively: x p 1 * = ( I A l s ) 1 u * and x p 2 * = ( I A l s 2 ) 1 ( A B u l / r + B u r / l ) .

3.3. Robust Data-Driven Feedback Controller Design

In order to ensure that the bipedal robot robustly stabilizes to a periodic orbit, we propose the following stepping controller:
u k R = u k * + K ( x k R x k * )
where x k * and u k * are the target state and step length at step k solved by (7) in the P1-Orbit or (8) in the P2-Orbit, x k R and u k R are the pre-impact state and step length of the current step, and K R 2 is the feedback gain to make the system stable.
Theorem 1.
When the closed-loop matrix A * + B * K is Schur-stable, the controller (9) ensures that the system converges to a disturbance invariant set E.
Proof. 
The system obtained in (6) can be written as
x k + 1 l s = A l s x k l s + B l s u k l s = [ A l s B l s ] x k l s u k l s
= ( [ A * , B * ] + W 0 X 0 U 0 ) x k l s u k l s
The uncertainty model defined as in (5) can be represented as
x k + 1 R = A * x k R + B * u k R + w k
Setting e k = x k + 1 R x k + 1 l s , substituting (9) into (12), and taking the difference with (11), we obtain the error dynamics:
e k + 1 = ( A * + B * K ) e k W 0 X 0 U 0 x k l s u k l s + w k
Since x k , u k , and w k are bounded in bipedal walking,
W 0 X 0 U 0 x k l s u k l s + w k Ω
If A * + B * K is Schur-stable, the error dynamics has a minimum disturbance invariant set E:
( A * + B * K ) E Ω E
Theorem 2.
For data given by X 0 , U 0 , X 1 in (2), if matrix X 0 U 0 has full rank, the Schur stability of A * + B * K equivalent to the following SDP problem is solvable,
f i n d J ,   P = P T 0 s . t . P D 11 0 D 12 0 P [ P J T ] D 21 P J D 22 0
where we define
D 11 D 12 D 21 D 22 : = X 1 X 1 T Δ Δ T X 1 X 0 U 0 T X 0 U 0 X 1 T X 0 U 0 X 0 U 0 T
and the feedback gain is K = J P 1 .
Proof. 
See Theorem 1 in reference [21]. □

4. Trajectory Generation and Whole-Body Control

We can generate the next-step-length input u k at each pre-impact state on both the x and y axes using (12). The desired swing foot trajectory x f R 3 , x ˙ f R 3 can be obtained by the fifth-order polynomial interpolation [24]. The desired CoM trajectory x C o M R 3 , x ˙ C o M R 3 and torso angular trajectory R t o r s o R 3 × 3 , ω t o r s o R 3 are specified by the user. The yaw angle of the swing leg is consistent with the torso, and the pitch and roll directions remain constant at zero, resulting in the swing foot angular trajectory R f o o t R 3 × 3 , ω f o o t R 3 . The reference acceleration in the task space for a linear/angular task is then tracked using a PD control, given by
x ¨ l i n d e s = K p ( x i r e f x i ) + K d ( x ˙ i r e f x ˙ i )
x ¨ a n g d e s = K p l o g ( R i T R i r e f ) + K d ( ω i r e f ω i )
where the x r e f and x corresponds to the i-th task reference trajectories and state estimates. K p and K d R 3 × 3 represent the PD gain matrix for different tasks.
The generated task-space trajectories are tracked using whole-body control. The specific QP formulation is as follows:
min q ¨ , f j i = 1 N t | | J i q ¨ + J ˙ i q ˙ x ¨ i d e s | | W i 2 + j = 1 N c | | f j | | W f 2
s . t . S f ( H q ¨ + C q ˙ + G j = 1 N c J c j T f j ) = 0
f j U
| q ¨ |     q ¨ m a x
where J i is the i-th task Jacobian and N t is the number of tasks, concluding the swing position/orientation task, torso orientation task, linear/angular moment, task and stance contact task. S f R 6 × 16 is the selection matrix associated with the underactuated DoF. The weight of the i-th task is set to W i R 3 × 3 . The decision variables q ¨ R 16 and f j R 6 are the joint acceleration and contact wrench. Specifically, for the stance contact constraint, J c j q ¨ + J ˙ c j q ˙ = 0 is also set as a term in the cost function in the form of a soft constraint. The rate of change of the centroidal angular momentum is defined as k ˙ d e s = K d k to suppress excessive angular momentum, and the centroidal momentum matrix [5] is used as the task Jacobian.
f j is the 6D force wrench for the j-th contact foot. For BRUCE’s line-foot, we use the contact wrench cone (CWC) described in [26] as the constraint in (20c).
By solving the QP problem, we obtain the optimal solutions q ¨ * and f j * , which can be used to calculate the joint torques through the joint dynamics of the system:
τ * = S a ( H q ¨ * + C q ˙ + G j = 1 N c J c j T f j * )
where S a R 10 × 16 is the selection matrix associated with the actuated DoF.

5. Results

5.1. Data Collection, Parameter Selection, and Feedback Gain

The collection of Hankel matrices in Equation (2) was conducted in the Gazebo simulator using a standard HLIP-based stepping controller [12], performing level walking in decoupled lateral and sagittal directions. For each direction, T s a m p l e = 1500 steps were collected to ensure persistence of excitation [18], with random assignments of desired velocity controlled via keyboard input within a speed range of [ 0.15 , 0.15 ] m/s.
The robustness of the stepping controller depends on the coverage range of the set C in (5). For the dataset collected from level walking, where the bounded disturbances of both linear and nonlinear systems are unknown, we roughly set a range [ d 1 / 2 m i n , d 1 / 2 m a x ] T corresponding to each component of Δ in Equation (4).
To enhance the stability of the feedback controller over a broader range of [ A * , B * ] , we introduced bounded Gaussian noise [ δ ω 1 , δ ω 2 ] T [ N ( μ 1 , σ 1 2 ) , N ( μ 2 , σ 2 2 ) ] T into the original state data X 1 , thereby increasing uncertainty, where the δ ω 1 / 2 [ δ ω 1 / 2 m i n , δ ω 1 / 2 m a x ] . This proved effective in practical simulations. Then, we noticed that each component of Δ should increase the upper bound of this Gaussian noise. In practical systems, excessively large noise bounds can render SDP infeasible; thus, we ultimately set the parameters as [ d 1 m i n , d 1 m a x ] = [ 0.002 , 0.002 ] , [ d 2 m i n , d 2 m a x ] = [ 0.01 , 0.01 ] , [ δ ω 1 m i n , δ ω 1 m a x ] = [ 0.004 , 0.004 ] and [ δ ω 1 m i n , δ ω 1 m a x ] = [ 0.006 , 0.006 ] . Therefore, each component of Δ ˜ with the introduced noise is equal to the matrix formed by the addition of the corresponding components of both.
Remark 2.
We use the original dataset D o r i g i n without introducing additional noise to calculate the target state in order to prevent the invariant disturbance set from expanding. The dataset D n o i s e with introduced noise, along with Δ ˜ Δ ˜ T , is used to calculate the gain K, stabilizing the system under larger C to counteract additional disturbances due to model uncertainties.
The robust feedback gain K is then determined through offline solving (16) using CVXPY [27]. The WBC in (20d) is solved in real time with 500 Hz using OSQP [28]. Notably, the SDP problem is solved offline, and the obtained feedback gain K is used to synthesize a deadbeat-style online stepping controller (9), ensuring real-time performance.
Remark 3.
Theoretically, the larger the added noise, the greater the maximum uncertainty that the stepping controller can stabilize, provided that Problem ( 16 ) remains feasible. However, excessive noise can render Problem ( 16 ) infeasible. Therefore, in our experiments, we selected the maximum noise that ensures the feasibility of Problem ( 16 ) as a parameter. Since the data collected from the simulator can be approximated as noise-free, the magnitude of the added bounded energy ellipsoidal noise serves as the boundary for the maximum disturbance that the system can encounter during experiments.

5.2. Simulation Results

5.2.1. Heavy Payload Task

The most direct manifestation of model uncertainty lies in the adaptability to different mass and inertia. From this perspective, the addition of noise in data collection is similar to the neighborhood adaptation concept in the field of reinforcement learning [29]. We hope that the stepping controller can adapt to a specific amount of payload without the need to re-collect a large amount of data for that specific payload. This situation can be considered as an uncertainty in the linear system [ A * , B * ] , which is uniformly stabilized through our method.
Figure 4 illustrates the adaptability of different algorithms on BRUCE when increasing its mass to alter its overall inertia. For comparison, data-driven control under the assumption of no noise [16] is also considered. To ensure statistical reliability, each algorithm runs 30 times in different scenarios. When the mass variation is small, HLIP and noise-free DDC can both ensure a certain success rate in 20 m walking. However, as the mass increases beyond 1 kg (approximately one-fifth of the entire mass of BRUCE), the robustness of our method allows the controller to overcome disturbances caused by changes in model parameters.

5.2.2. Sloping Terrain Task

Walking on slopes introduces acceleration z ¨ C o M in BRUCE’s vertical direction, thereby altering the S2S dynamics. We did not make any special adjustments to the swing leg trajectory generation, causing early touchdown when the desired speed increases.
Figure 5 shows the maximum walking speeds of the RDDC and HLIP stepping controllers under different slopes. As velocity increases, vertical acceleration in the z-direction also increases, leading to a shorter swing time T s s p . The simulation results demonstrate that RDDC exhibits better robustness under the influence of two types of uncertainty parameters.

5.2.3. Push Recovery Task

In this section, we compare the lateral and lateral disturbance rejection capabilities of the HLIP/RDDC-based stepping controller. In simulation experiments, BRUCE tracked the same variable velocity command. The CoM offsets of our method and HLIP initial controller under different directions and external forces are shown in the Figure 6 and Figure 7. External forces were applied at the beginning of each step cycle for the same duration t force = 20 ms , approximately one-tenth of a step cycle, simulating the impact of external forces on the actual system.
Remark 4.
The HLIP-based method [12] can achieve locomotion capabilities similar to RDDC during BRUCE’s walking on flat ground. However, due to the larger ratio of leg inertia to total body inertia in BRUCE compared to Cassie [12], it is more difficult to use LIP-like models to capture key physical characteristics. The DDC method, to some extent, achieves a more accurate linear system. At the same time, the robust approach based on ellipsoidal noise captures the effects of more general model uncertainties. Therefore, in experiments across various environments, our method outperforms the HLIP method.

5.2.4. Comparison with Indirect Data-Driven Methods

Unlike direct data-driven methods, indirect data-driven control [14] typically involves two steps: offline parameter identification to obtain the system dynamics parameters, followed by computation of the feedback gain K. The intuitive procedural differences between the two approaches are illustrated in Figure 8 below.

5.2.5. Sensitivity Analysis Under Different Noise Levels

We can solve for the stable control gain K under the given range of uncertain noise (4) using Equation (16). Its stability can be naturally obtained through the Lyapunov stability condition for discrete systems [21].
To verify the maximum noise level that ellipsoids of different noise levels can withstand, we added various levels of noise to the collected original data. Correspondingly, the size of the bounded energy ellipsoid also changed. We compared the maximum noise levels that can be resisted under the HLIP model for different gains K derived by solving the SDP problem constructed with bounded energy ellipsoids after adding different noise levels. The corresponding results are shown in Figure 9 below.
By directly testing on the HLIP, we can clearly observe the response changes under different disturbances, demonstrating the sensitivity of bounded energy ellipsoids of varying sizes to perturbations. We observed that adjusting the size of the bounded energy ellipsoid helps to withstand larger disturbances. Although the improvement is modest, it has been shown to enhance the robustness of the walking controller in the actual BRUCE simulation.

5.2.6. Sim-to-Sim Experiments

We also conducted sim-to-sim experiments from Gazebo to Mujoco, calculating feedback gains from data collected under different lengths. First, walking was performed in Gazebo using a nominal controller (HLIP), and data were collected for 500 steps (−0.1 m/s to −0.1 m/s), 1000 steps (−0.2 m/s to −0.2 m/s), and 3000 steps (−0.2 m/s to −0.2 m/s). We observed that stable walking in Mujoco can only be achieved when a larger amount of data is collected and broader range of walking speeds is covered.
The feedback gains obtained by solving (16) differs based on the length of the collected data, as shown in Table 2 below; refer to Figure 10. We increased the complexity of the walking scenario in Mujoco by adding a 0.02 m step to the ground. When the amount of collected data is small, the controller causes a fall after walking for some time. However, with a larger data sample, BRUCE can successfully traverse this terrain.

6. Conclusions and Future Work

In this paper, we directly construct a state-feedback controller using input-state data from S2S dynamics during bipedal walking, along with an additional noise sequence. Given the range of uncertainties, we can design robust feedback gains to stabilize a family of linear systems. Our work not only demonstrates the potential of behavior system theory in gait control for legged robots, but also shows that this data-driven control method, compared to other related data-driven approaches such as deep reinforcement learning or supervised learning, does not require large-scale data and training costs, achieving robustness similar to or even better than model-based control methods. Moreover, we have developed a mathematically rigorous robust controller that theoretically characterizes the possible disturbances arising from model uncertainties. Unlike machine learning’s “black box”, this method provides a more comprehensive explanation for specific types of model uncertainties. At the same time, compared to robust model predictive control, this method does not require online solutions to robust optimal control problems (such as min–max optimal control), thereby improving computational real-time performance.
Through several representative simulation scenarios, the potential of behavior system theory-based data-driven control in humanoid locomotion control has been demonstrated.
However, the current method still has some limitations and challenges, which we will address in future work.
  • We found that the stepping controller based on RDDC does not show significant improvement in speed tracking compared to the HLIP-based controller. This is because we only used the ellipsoidal matrix center of C as the reference dynamics. When the uncertainty range of C is large, it results in a large error invariant set.
  • The S2S dynamics are inherently nonlinear. If a nonlinear mathematical structure is determined, a state-feedback controller can be constructed by feedback linearization of this system using [30].
  • The system behavior obtained by the nominal controller (HLIP or other stepping controllers) is limited. When walking at higher speeds, the nonlinear effect of the swing leg inertia increases. DeePO [31] can be used for online updating of the K value to address this issue.

Author Contributions

C.S. and X.Z. co-organized this work and wrote the manuscript draft. B.C. wrote some code. C.L., S.H. and Y.Z. proposed some good ideas. J.Z. supervised the research and commented on the manuscript draft. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Major Research Plan of the National Natural Science Foundation of China under Grant no. 92048301 and the National Outstanding Youth Science Fund Project of National Natural Science Foundation of China (Grant no. 52025054).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data and simulation code used to support the findings of this study are available from Chao Song upon request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Faraji, S.; Ijspeert, A.J. 3LP: A linear 3D-walking model including torso and swing dynamics. Int. J. Robot. Res. 2017, 36, 436–455. [Google Scholar] [CrossRef]
  2. Kajita, S.; Morisawa, M.; Miura, K.; Nakaoka, S.; Harada, K.; Kaneko, K.; Kanehiro, F.; Yokoi, K. Biped walking stabilization based on linear inverted pendulum tracking. In Proceedings of the 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems, Taipei, Taiwan, 18–22 October 2010; pp. 4489–4496. [Google Scholar]
  3. Khadiv, M.; Herzog, A.; Moosavian, S.A.A.; Righetti, L. Walking control based on step timing adaptation. IEEE Trans. Robot. 2020, 36, 629–643. [Google Scholar] [CrossRef]
  4. Lu, X.; Cannon, M. Robust adaptive model predictive control with persistent excitation conditions. Automatica 2023, 152, 110959. [Google Scholar] [CrossRef]
  5. Orin, D.E.; Goswami, A.; Lee, S.H. Centroidal dynamics of a humanoid robot. Auton. Robot. 2013, 35, 161–176. [Google Scholar] [CrossRef]
  6. Li, J.; Nguyen, Q. Dynamic walking of bipedal robots on uneven stepping stones via adaptive-frequency MPC. IEEE Control Syst. Lett. 2023, 7, 1279–1284. [Google Scholar] [CrossRef]
  7. Dafarra, S.; Romualdi, G.; Pucci, D. Dynamic complementarity conditions and whole-body trajectory optimization for humanoid robot locomotion. IEEE Trans. Robot. 2022, 38, 3414–3433. [Google Scholar] [CrossRef]
  8. Caron, S.; Escande, A.; Lanari, L.; Mallein, B. Capturability-Based Pattern Generation for Walking with Variable Height. IEEE Trans. Robot. 2020, 36, 517–536. [Google Scholar] [CrossRef]
  9. Mihalec, M.; Yi, J. Balance Gait Controller for a Bipedal Robotic Walker with Foot Slip. IEEE/ASME Trans. Mechatron. 2023, 28, 2012–2019. [Google Scholar] [CrossRef]
  10. Ding, J.; Han, L.; Ge, L.; Liu, Y.; Pang, J. Robust Locomotion Exploiting Multiple Balance Strategies: An Observer-Based Cascaded Model Predictive Control Approach. IEEE/ASME Trans. Mechatron. 2022, 27, 2089–2097. [Google Scholar] [CrossRef]
  11. Jeong, H.; Lee, I.; Oh, J.; Lee, K.K.; Oh, J.H. A Robust Walking Controller Based on Online Optimization of Ankle, Hip, and Stepping Strategies. IEEE Trans. Robot. 2019, 35, 1367–1386. [Google Scholar] [CrossRef]
  12. Xiong, X.; Ames, A. 3-d underactuated bipedal walking via h-lip based gait synthesis and stepping stabilization. IEEE Trans. Robot. 2022, 38, 2405–2425. [Google Scholar] [CrossRef]
  13. Ames, A.D.; Galloway, K.; Sreenath, K.; Grizzle, J.W. Rapidly exponentially stabilizing control lyapunov functions and hybrid zero dynamics. IEEE Trans. Autom. Control 2014, 59, 876–891. [Google Scholar] [CrossRef]
  14. Xiong, X.; Chen, Y.; Ames, A.D. Robust disturbance rejection for robotic bipedal walking: System-level-synthesis with step-to-step dynamics approximation. In Proceedings of the 2021 60th IEEE Conference on Decision and Control (CDC), Austin, TX, USA, 14–17 December 2021; pp. 697–704. [Google Scholar]
  15. Hinojosa, E.H.; Bhounsule, P.A. Data-driven Identification of a Non-homogeneous Inverted Pendulum Model for Enhanced Humanoid Control. In Proceedings of the 2023 IEEE-RAS 22nd International Conference on Humanoid Robots (Humanoids), Austin, TX, USA, 12–14 December 2023; pp. 1–8. [Google Scholar]
  16. De Persis, C.; Tesi, P. Formulas for data-driven control: Stabilization, optimality, and robustness. IEEE Trans. Autom. Control 2019, 65, 909–924. [Google Scholar] [CrossRef]
  17. Dai, M.; Xiong, X.; Ames, A.D. Data-driven step-to-step dynamics based adaptive control for robust and versatile underactuated bipedal robotic walking. arXiv 2022, arXiv:2209.08458. [Google Scholar]
  18. Willems, J.C.; Rapisarda, P.; Markovsky, I.; De Moor, B.L. A note on persistency of excitation. Syst. Control Lett. 2005, 54, 325–329. [Google Scholar] [CrossRef]
  19. Markovsky, I.; Dörfler, F. Behavioral systems theory in data-driven analysis, signal processing, and control. Annu. Rev. Control 2021, 52, 42–64. [Google Scholar] [CrossRef]
  20. van Waarde, H.J.; Camlibel, M.K.; Mesbahi, M. From noisy data to feedback controllers: Nonconservative design via a matrix S-lemma. IEEE Trans. Autom. Control 2020, 67, 162–175. [Google Scholar] [CrossRef]
  21. Bisoffi, A.; De Persis, C.; Tesi, P. Data-driven control via Petersen’s lemma. Automatica 2022, 145, 110537. [Google Scholar] [CrossRef]
  22. Bisoffi, A.; De Persis, C.; Tesi, P. Learning controllers for performance through LMI regions. IEEE Trans. Autom. Control 2022, 68, 4351–4358. [Google Scholar] [CrossRef]
  23. Berberich, J.; Scherer, C.W.; Allgöwer, F. Combining prior knowledge and data for robust controller design. IEEE Trans. Autom. Control 2022, 68, 4618–4633. [Google Scholar] [CrossRef]
  24. Shen, J.; Zhang, J.; Liu, Y.; Hong, D. Implementation of a robust dynamic walking controller on a miniature bipedal robot with proprioceptive actuation. In Proceedings of the 2022 IEEE-RAS 21st International Conference on Humanoid Robots (Humanoids), Ginowan, Japan, 28–30 November 2022; pp. 39–46. [Google Scholar]
  25. Gibson, G.; Dosunmu-Ogunbi, O.; Gong, Y.; Grizzle, J. Terrain-adaptive, alip-based bipedal locomotion controller via model predictive control and virtual constraints. In Proceedings of the 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Kyoto, Japan, 23–27 October 2022; pp. 6724–6731. [Google Scholar]
  26. Ding, Y.; Khazoom, C.; Chignoli, M.; Kim, S. Orientation-aware model predictive control with footstep adaptation for dynamic humanoid walking. In Proceedings of the 2022 IEEE-RAS 21st International Conference on Humanoid Robots (Humanoids), Ginowan, Japan, 28–30 November 2022; pp. 299–305. [Google Scholar]
  27. Agrawal, A.; Verschueren, R.; Diamond, S.; Boyd, S. A rewriting system for convex optimization problems. J. Control Decis. 2018, 5, 42–60. [Google Scholar] [CrossRef]
  28. Stellato, B.; Banjac, G.; Goulart, P.; Bemporad, A.; Boyd, S. OSQP: An operator splitting solver for quadratic programs. Math. Program. Comput. 2020, 12, 637–672. [Google Scholar] [CrossRef]
  29. Jenelten, F.; He, J.; Farshidian, F.; Hutter, M. Dtc: Deep tracking control. Sci. Robot. 2024, 9, eadh5401. [Google Scholar] [PubMed]
  30. De Persis, C.; Rotulo, M.; Tesi, P. Learning controllers from data via approximate nonlinearity cancellation. IEEE Trans. Autom. Control 2023, 68, 6082–6097. [Google Scholar] [CrossRef]
  31. Zhao, F.; Dörfler, F.; Chiuso, A.; You, K. Data-Enabled Policy Optimization for Direct Adaptive Learning of the LQR. arXiv 2024, arXiv:2401.14871. [Google Scholar]
Figure 1. BRUCE carrying payload and walking up a slope.
Figure 1. BRUCE carrying payload and walking up a slope.
Biomimetics 09 00681 g001
Figure 2. Configuration of BRUCE robot.
Figure 2. Configuration of BRUCE robot.
Biomimetics 09 00681 g002
Figure 3. The proposed control framework.
Figure 3. The proposed control framework.
Biomimetics 09 00681 g003
Figure 4. The success rate of walking forward for 20 m with different stepping controllers.
Figure 4. The success rate of walking forward for 20 m with different stepping controllers.
Biomimetics 09 00681 g004
Figure 5. A heat map of maximum walking speeds under varying slopes: (a) HLIP stepping controller; (b) RDDC stepping controller.
Figure 5. A heat map of maximum walking speeds under varying slopes: (a) HLIP stepping controller; (b) RDDC stepping controller.
Biomimetics 09 00681 g005
Figure 6. Push recovery of 30 N in X-direction: (a) HLIP stepping controller; (b) RDDC stepping controller.
Figure 6. Push recovery of 30 N in X-direction: (a) HLIP stepping controller; (b) RDDC stepping controller.
Biomimetics 09 00681 g006
Figure 7. Push recovery of 45 N in Y-direction: (a) HLIP stepping controller; (b) RDDC stepping controller.
Figure 7. Push recovery of 45 N in Y-direction: (a) HLIP stepping controller; (b) RDDC stepping controller.
Biomimetics 09 00681 g007
Figure 8. The differences between the direct and indirect data-driven control frameworks are as follows. In direct data-driven control, the matrix family C is used only as an intermediate parameter, represented by bounded energy ellipsoidal parameterization, and is not solved explicitly. This represents a fundamental difference from indirect data-driven control.
Figure 8. The differences between the direct and indirect data-driven control frameworks are as follows. In direct data-driven control, the matrix family C is used only as an intermediate parameter, represented by bounded energy ellipsoidal parameterization, and is not solved explicitly. This represents a fundamental difference from indirect data-driven control.
Biomimetics 09 00681 g008
Figure 9. The results in figures (ac) show the velocity responses of the HLIP after walking within the same speed range, subjected to disturbances with added noise ranges with absolute values 0.04, 0.042, and 0.043, respectively. The ellipsoid energy level matches the noise range in each case. Figure (d) illustrates that when the noise range is 0.04, the system diverges under disturbances of the corresponding magnitude.
Figure 9. The results in figures (ac) show the velocity responses of the HLIP after walking within the same speed range, subjected to disturbances with added noise ranges with absolute values 0.04, 0.042, and 0.043, respectively. The ellipsoid energy level matches the noise range in each case. Figure (d) illustrates that when the noise range is 0.04, the system diverges under disturbances of the corresponding magnitude.
Biomimetics 09 00681 g009
Figure 10. Sim-to-sim experiments in Mujoco using a stepping controller built from Gazebo data of varying lengths: (a) 500 steps, (b) 1000 steps, (c) 3000 steps.The red circle and cross marks represent successful passes and failures due to falls, respectively.
Figure 10. Sim-to-sim experiments in Mujoco using a stepping controller built from Gazebo data of varying lengths: (a) 500 steps, (b) 1000 steps, (c) 3000 steps.The red circle and cross marks represent successful passes and failures due to falls, respectively.
Biomimetics 09 00681 g010
Table 1. Physical parameters of robot.
Table 1. Physical parameters of robot.
ParameterSymbolValue
Torso Mass (kg) M t o r s o 1.31
Each Leg Mass (kg) M l e g 1.44
Torso Inertia (kg·m2) I x x / I y y / I z z 0.013 / 0.010 / 0.005
Leg Inertia (kg·m2) I x x / I y y / I z z 0.021 / 0.020 / 0.001
Rate of Mass M l e g / M t o r s o 1.09
Table 2. Feedback gain of different lengths of collected data.
Table 2. Feedback gain of different lengths of collected data.
Number of StepsFeedback GainSuccess or Failure
500 [ 1.075 , 0.43 ] Failure
1000 [ 1.205 , 0.38 ] Failure
3000 [ 1.375 , 0.26 ] Success
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Song, C.; Zang, X.; Chen, B.; Heng, S.; Li, C.; Zhu, Y.; Zhao, J. Bipedal Stepping Controller Design Considering Model Uncertainty: A Data-Driven Perspective. Biomimetics 2024, 9, 681. https://doi.org/10.3390/biomimetics9110681

AMA Style

Song C, Zang X, Chen B, Heng S, Li C, Zhu Y, Zhao J. Bipedal Stepping Controller Design Considering Model Uncertainty: A Data-Driven Perspective. Biomimetics. 2024; 9(11):681. https://doi.org/10.3390/biomimetics9110681

Chicago/Turabian Style

Song, Chao, Xizhe Zang, Boyang Chen, Shuai Heng, Changle Li, Yanhe Zhu, and Jie Zhao. 2024. "Bipedal Stepping Controller Design Considering Model Uncertainty: A Data-Driven Perspective" Biomimetics 9, no. 11: 681. https://doi.org/10.3390/biomimetics9110681

APA Style

Song, C., Zang, X., Chen, B., Heng, S., Li, C., Zhu, Y., & Zhao, J. (2024). Bipedal Stepping Controller Design Considering Model Uncertainty: A Data-Driven Perspective. Biomimetics, 9(11), 681. https://doi.org/10.3390/biomimetics9110681

Article Metrics

Back to TopTop