This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Whole-Body Dynamic Telelocomotion: A Step-to-Step Dynamics Approach to Human Walking Reference Generation

Guillermo Colin1, Joseph Byrnes2, Youngwoo Sim1, Patrick M. Wensing3, and Joao Ramos1 This work is supported by the National Science Foundation via grant CMMI-2043339.1 Department of Mechanical Science and Engineering at the University of Illinois at Urbana-Champaign, USA.(gjcolin3@illinois.edu)2 Department of Industrial and Enterprise Systems Engineering at the University of Illinois at Urbana-Champaign, USA.3 Department of Aerospace and Mechanical Engineering at the University of Notre Dame, USA.
Abstract

Teleoperated humanoid robots hold significant potential as physical avatars for humans in hazardous and inaccessible environments, with the goal of channeling human intelligence and sensorimotor skills through these robotic counterparts. Precise coordination between humans and robots is crucial for accomplishing whole-body behaviors involving locomotion and manipulation. To progress successfully, dynamic synchronization between humans and humanoid robots must be achieved. This work enhances advancements in whole-body dynamic telelocomotion, addressing challenges in robustness. By embedding the hybrid and underactuated nature of bipedal walking into a virtual human walking interface, we achieve dynamically consistent walking gait generation. Additionally, we integrate a reactive robot controller into a whole-body dynamic telelocomotion framework. Thus, allowing the realization of telelocomotion behaviors on the full-body dynamics of a bipedal robot. Real-time telelocomotion simulation experiments validate the effectiveness of our methods, demonstrating that a trained human pilot can dynamically synchronize with a simulated bipedal robot, achieving sustained locomotion, controlling walking speeds within the range of 0.0 m/s to 0.3 m/s, and enabling backward walking for distances of up to 2.0 m. This research contributes to advancing teleoperated humanoid robots and paves the way for future developments in synchronized locomotion between humans and bipedal robots.

I Introduction

The full potential of teleoperated humanoid robots has yet to be realized, particularly in their role as physical avatars for humans in hazardous or inaccessible locations [1]. Although noteworthy achievements have been made in loco-manipulation tasks, performance remains limited [2, 3]. This is because for loco-manipulation task completion precise coordination of whole-body dynamics is essential [4, 5]. As a result, in teleoperated humanoid robot research a major obstacle is the lack of synchronized locomotion between human and robot, which entails coordinated movements like simultaneous lifting-off and stepping-down. The disparity in locomotion dynamics between human and robot makes it challenging to fully leverage human capability. Hence, establishing a dynamic link between human and robot locomotion is crucial. In recent years, efforts have been made to realize this critical behavior [6, 7, 3]. However, this area of research remains largely unexplored, particularly within the bipedal locomotion control community.

Refer to caption
Figure 1: The human pilot (right) uses our human-machine interface to dynamically synchronize with the simulated bipedal robot Tello (left) using our whole-body dynamic telelocomotion framework. Supporting video link: https://youtu.be/_ecsf-QxVYg

Synchronizing the gait in real-time between a human and a bipedal robot poses considerable challenges. First, while humans and humanoid robots have a similar morphology, they have fundamental differences. This motivates the need to transfer locomotion strategies from humans to robots. One approach is to abstract human locomotion through reduced-order models [8, 9]. The second challenge is that the human and robot have walking dynamics that are destined to evolve at different spatial and temporal scales. This is primarily due to the underactuated nature of walking coupled with the difference in natural frequencies between human and robot. This motivates the use of bilateral feedback between human and robot [9, 10, 11]. In order to inform the human pilot of robot dynamics, haptic force feedback can encode information about the robot’s sense of balance, i.e., stability, in an instinctual manner to harness human sensorimotor skills. The final challenge is the development of specialized hardware. This is why human-machine interfaces that facilitate whole-body motion capture and haptic force feedback have been instrumental in advancing teleoperated humanoid robot research [12, 13, 14].

Encouraging results have been obtained from the unification of three concepts: abstraction of humanoid locomotion, scaling of human whole-body motion to robot proportions, and dynamic synchronization of human and robot via bilateral feedback. This framework, termed whole-body dynamic telelocomotion, has demonstrated synchronized balancing and stepping between a human pilot and bipedal robot in hardware [6]. In our previous work, it yielded synchronized dynamic walking of a simulated bipedal robot [15]. The extension from stepping to walking was enabled by a virtual interface linking human and bipedal robot. This interface serves as a virtual reference of human walking behavior generated from their stepping motion. Although similar methods have been proposed, [16, 17], they lacked integration into a holistic framework to enable the transfer of intelligence and sensorimotor skills from human to robot.

While the results of [15] were promising, sustained synchronized locomotion between human and robot was challenging. Consistently reaching more than 1.0 m of robot walking through teleoperation was difficult. The human pilot could generate a walking gait through a step, but struggled to generate transitions between walking gaits. Continuously adapting to robot walking dynamics proved taxing over time. On the other hand, bipedal robot walking using feedback control methods has become highly robust [18, 19]. These control methods benefit from considering the hybrid and underactuated dynamics of walking, allowing seamless transitions between walking gaits, particularly at different speeds.

In this work, the hybrid linear inverted pendulum model (H-LIPM) [20] is utilized to abstract our human walking reference and thus capture the hybrid and underactuated dynamics of walking within our whole-body dynamic telelocomotion framework. The same mapping of human stepping motion to a desired walking gait is utilized as in [15] to infer human locomotion strategy. This allows the human pilot to intuitively update the H-LIPM dynamics during telelocomotion. Additionally, a reactive robot controller is designed to reproduce scaled human motion on the robot’s full-body dynamics. The main contribution of this work lies in incorporating the step-to-step dynamics of walking into our human walking reference interface. The result is dynamically consistent human walking reference motion. Another contribution is the integration of a reactive robot controller into our telelocomotion framework. As depicted in Fig. 1, the presented methods are validated through robot full-body dynamics simulation experiments conducted with our telelocomotion framework. The results demonstrate that a human pilot can dynamically synchronize with the humanoid robot Tello to achieve sustained locomotion. This is demonstrated by the pilot’s capacity to control robot walking speeds within the range of 0.0 m/s to 0.3 m/s, and maintain robot backward walking for distances of up to 2.0 m.

II Human Walking Reference Generation

The goal of incorporating step-to-step dynamics into our human walking reference (HWR) interface is to facilitate dynamically consistent HWR motion generation by the human pilot. The proposed methodology is presented in this section and summarized in Fig. 2. First, a theoretical basis for creating an abstraction of the HWR using a reduced-order model is presented, namely the hybrid linear inverted pendulum model (H-LIPM). Next, a method from [15] of mapping a human’s stepping motion to their desired walking behavior is reviewed. Finally, we highlight the major contribution from this work and detail the HWR motion generation algorithm.

Refer to caption
Figure 2: The human pilot (a) transitions between desired walking gaits (b) by stepping along the human-machine interface. These gaits are generated using a mapping that translates human stepping into a period 1 orbit. To achieve the target walking behavior, a deadbeat controller is utilized on the human walking reference dynamics, abstracted using the H-LIPM reduced-order model (d). This approach embeds the step-to-step dynamics of walking (e) into our human walking reference motion generation process.

II-A H-LIP Model

The H-LIPM [21], an extension of the LIPM, is a two-domain hybrid system that encapsulates the hybrid dynamics of point-foot walking. The two domains of walking are a single support phase (SSP) and a double support phase (DSP). The domain-specific continuous dynamics of the H-LIPM are governed by

SSP: x¨=ω2x,\displaystyle\ddot{x}=\omega^{2}x, (1)
DSP: x¨=0,\displaystyle\ddot{x}=0,

where xx is the center of mass (CoM) position relative to the stance foot, ω=gh\omega=\sqrt{\frac{g}{h}} is the natural frequency under the gravity field gg, and hh is the CoM height. The H-LIPM describes the discrete transitions between SSP (S) and DSP (D) through smooth hybrid system reset maps, which are defined as follows:

ΔSD:\displaystyle\Delta_{S\rightarrow D}: {x+=xx˙+=x˙,\displaystyle\begin{cases}x^{+}=x^{-}\\ \dot{x}^{+}=\dot{x}^{-}\end{cases}, (2a)
ΔDS:\displaystyle\Delta_{D\rightarrow S}: {x+=xx˙+=x˙,\displaystyle\begin{cases}x^{+}=x^{-}-\ell\\ \dot{x}^{+}=\dot{x}^{-}\end{cases}, (2b)

where the control input \ell is the step length and the superscripts ++ and - denote CoM states after and before each discrete state reset, respectively. The closed-form solution of the hybrid system is derived by solving Eqs. (1) - (2). This solution defines the time evolution of the H-LIPM dynamics, as shown below:

SSP: {x(t)=c1eωt+c2eωtx˙(t)=ω(c1eωtc2eωt),\displaystyle\begin{cases}x(t)=c_{1}e^{\omega t}+c_{2}e^{-\omega t}\\ \dot{x}(t)=\omega(c_{1}e^{\omega t}-c_{2}e^{-\omega t})\end{cases}, (3a)
DSP: {x(t)=x+x˙tx˙(t)=x˙,\displaystyle\begin{cases}x(t)=x^{-}+\dot{x}^{-}t\\ \dot{x}(t)=\dot{x}^{-}\end{cases}, (3b)

where time in each phase is denoted by t[0,TSSP]t\in[0,T_{SSP}] or t[0,TDSP]t\in[0,T_{DSP}]. TSSPT_{SSP} and TDSPT_{DSP} are the duration of SSP and DSP, respectively. The constants c1=12(x++1ωx˙+)c_{1}=\frac{1}{2}(x^{+}+\frac{1}{\omega}\dot{x}^{+}) and c2=12(x+1ωx˙+)c_{2}=\frac{1}{2}(x^{+}-\frac{1}{\omega}\dot{x}^{+}) are constants of integration.

II-B Human Stepping Motion Mapping

A mapping of human stepping motion to desired walking motion was introduced in [15]. The two assumptions made are that 1) the desired human walking gait corresponds to a stable walking gait of the canonical LIPM and 2) the human pilot’s (h) stepping frequency and CoM position along our human-machine interface (HMI) are sufficient for the pilot to intuitively generate these HWR (h¯\overline{\textsc{h}}) gaits.

From a technical perspective, these two assumptions allow mapping the human pilot’s stepping frequency, Ts1T_{s}^{-1}, and CoM position, xhx_{\textsc{h}}, to a stable walking gait. The walking gait chosen to represent nominal human walking is characterized as satisfying (x+,x˙+)=(x,x˙)\left({x}^{+},\dot{x}^{+}\right)=\left(-{x}^{-},\dot{x}^{-}\right) during a single step. This one-step periodic orbit, known as a P1 orbit, can be generated in real-time from TsT_{s} and xhx_{\textsc{h}}. The core idea is to map xhξxh¯x_{\textsc{h}}\rightarrow\xi_{x\overline{\textsc{h}}}^{-}, i.e., the human CoM position along the HMI equates to some desired pre-impact (end-of-step) value of the walking reference’s divergent component of motion (DCM). The DCM, defined ξx:=x+x˙ω\xi_{x}:=x+\frac{\dot{x}}{\omega}, is the key state tracked in our telelocomotion framework to synchronize human and robot locomotion as it captures the unstable component of motion of the pendulum dynamics [9].

It can be shown that ξx\xi_{x}^{-} and TsT_{s} can completely characterize a P1 orbit given some necessary and sufficient conditions are satisfied. Full details can be found in [15]. In general, from human stepping motion, the desired pre-impact CoM states of a stable walking gait can be computed as following:

{xh¯(Ts,xh)=xh/(1+σ1hωh),x˙h¯(Ts,xh)=ωh(xhxh¯),\begin{cases}x_{\overline{\textsc{h}}}^{-}(T_{s},x_{\textsc{h}})=x_{\textsc{h}}/({1+\frac{\sigma_{1\textsc{h}}}{\omega_{\textsc{h}}}}),\\ \dot{x}_{\overline{\textsc{h}}}^{-}(T_{s},x_{\textsc{h}})=\omega_{\textsc{h}}(x_{\textsc{h}}-x_{\overline{\textsc{h}}}^{-}),\end{cases} (4)

where σ1h\sigma_{1\textsc{h}} is the orbital slope of the human LIPM, defined σ1h:=ωhcoth(Ts2ωhmissing).\sigma_{1\textsc{h}}:=\omega_{\textsc{h}}\>\coth\Big(\frac{T_{s}}{2}\omega_{\textsc{h}}\Big{missing}). We can extend this mapping to stable walking gaits of the H-LIPM.

II-C S2S Dynamics Approach

The key idea is to abstract the HWR as a H-LIPM with a control policy that captures the desired human pilot’s walking behavior while obeying the step-to-step dynamics of walking. In our telelocomotion framework, this will enhance the robot controller’s capability to consistently track the motion of the HWR when scaled to robot proportions.

The approach in [15] forced the pilot to learn the robot’s dynamics of transitioning between different stable walking gaits. The HWR was assumed to be a LIPM whose dynamics always evolved along a P1 orbit. In order to aid the pilot, a haptic virtual spring was implemented to give them an acceleration sensation similar to that of the HWR. This allowed for successful HWR motion generation and dynamic synchronized walking between human and bipedal robot, but it was not sustainable for prolonged periods.

The non-trivial ability to traverse between these stable walking gaits can be more accurately explained by the step-to-step (S2S) dynamics of bipedal walking [20]. The S2S dynamics of walking are perfectly highlighted by the H-LIPM. As shown in Eq. (3), under the H-LIPM approximation, the dynamics of walking are purely passive throughout the continuous domains. From Eq. (2b), the control input, \ell, is applied only discretely. This motivates viewing the dynamics of walking at the step level. The result is a discrete linear dynamical system,

𝐱k+1=A𝐱k+Buk,\mathbf{x}_{k+1}=A\mathbf{x}_{k}+Bu_{k}, (5)

that maps step kk pre-impact CoM states, 𝐱k\mathbf{x}_{k}, to step k+1k+1 pre-impact CoM states, 𝐱k+1\mathbf{x}_{k+1}, via the step length uk=u_{k}=\ell. The state transition matrix, AA, and control input mapping vector, BB, are defined by combining Eqs. (2) - (3). Thus, stabilization of 𝐱\mathbf{x} to a desired 𝐱\mathbf{x}^{*} in Eq. (5) can be enforced with feedback control. The case of particular interest is when 𝐱\mathbf{x}^{*} corresponds to a P1 orbit. In this case, the deadbeat controller from [20] can optimally drive 𝐞:=(𝐱𝐱)𝟎\mathbf{e}:=(\mathbf{x}-\mathbf{x^{*}})\rightarrow\mathbf{0} in two steps:

uk=x+x+TDSPx˙+coth(TSSPω)ω(x˙x˙).u_{k}=x^{-}+x^{-*}+T_{DSP}\dot{x}^{-}+\frac{\coth(T_{SSP}\omega)}{\omega}(\dot{x}^{-}-\dot{x}^{-*}). (6)

In this work, Eq. (6) is the control policy of the HWR. The control objective is to achieve the desired human locomotion strategy captured by Eq. (4) on the dynamics of the HWR. The dynamics of the HWR evolve according to Eq. (3). By modeling the HWR as a H-LIPM with step-to-step dynamics given by Eq. (5), the hybrid and underactuated nature of walking is considered in the motion generation of the HWR as outlined in Algorithm 1.

Algorithm 1 HWR Motion Generation
1:Initialize (xh¯+x_{\overline{\textsc{h}}}^{+}, x˙h¯+\dot{x}_{\overline{\textsc{h}}}^{+}) & tt
2:while Telelocomotion do
3:    if SSPDSPSSP\rightarrow DSP \lor DSPSSPDSP\rightarrow SSP then
4:        Set (xh¯x_{\overline{\textsc{h}}}^{-}, x˙h¯\dot{x}_{\overline{\textsc{h}}}^{-}) = (xh¯x_{\overline{\textsc{h}}}, x˙h¯\dot{x}_{\overline{\textsc{h}}})
5:        if SSPDSPSSP\rightarrow DSP then
6:           Update (xh¯+x_{\overline{\textsc{h}}}^{+}, x˙h¯+\dot{x}_{\overline{\textsc{h}}}^{+}) with Eq. (2a)
7:        else if DSPSSPDSP\rightarrow SSP then
8:           Update (xh¯+x_{\overline{\textsc{h}}}^{+}, x˙h¯+\dot{x}_{\overline{\textsc{h}}}^{+}) with Eq. (2b)
9:        end if
10:        Reset t=0t=0
11:    else
12:        if SSPSSP then
13:           Update (xh¯x_{\overline{\textsc{h}}}, x˙h¯\dot{x}_{\overline{\textsc{h}}}) with Eq. (3a)
14:           Update 𝐱\mathbf{x}^{*} with Eq. (4)
15:           Predict 𝐱\mathbf{x} with Eq. (3a)
16:           Compute uku_{k} with Eq. (6)
17:        else if DSPDSP then
18:           Update (xh¯x_{\overline{\textsc{h}}}, x˙h¯\dot{x}_{\overline{\textsc{h}}}) with Eq. (3b)
19:        end if
20:    end if
21:end while

III Robot Controller

The other contribution of this work is the integration of a reactive robot controller into our whole-body dynamic telelocomotion framework to facilitate dynamic synchronized locomotion between human pilot and bipedal robot. The objective is to reactively reproduce scaled human motion and implement the human pilot’s locomotion strategy. This strategy is abstracted using reduced-order models. As a result, in our telelocomotion framework, dynamic synchronization between human and robot is enforced at the reduced-order model level. The role of the robot controller is to track the resulting robot 2D LIP trajectory at the full-body dynamics level. This trajectory represents stable locomotion (center of mass dynamics) and the corresponding control input (center of pressure strategy).

The robot controller considers the full-body dynamics of the humanoid robot Tello [22]. The controller consists of an optimization-based balance controller that computes joint torques for the legs in stance during DSP and SSP. As shown in Fig. 3, the balance controller approximates the robot full-body dynamics using the single rigid-body model (SRBM). The controller also consists of a swing-leg controller that computes joint torques for the leg in swing during SSP.

Refer to caption
Figure 3: The SRBM approximation of the full-body dynamics of the bipedal robot Tello (shown here during SSPSSP). The floating-base dynamics of the robot are actuated by ground reaction forces (GRFs) applied at active contact points. These contact points are located at the toe and heel of the robot’s feet.
Refer to caption
Figure 4: High-level overview of our whole-body dynamic telelocomotion framework. The sub-components discussed in this work are labeled with their respective sections for ease of navigation. Previous work is highlighted by the dashed colored lines. For example, the HMI (purple color) in [14] and dynamic synchronization via bilateral feedback teleoperation (green color) in [15].

III-A Robot Model

The 15.8 kg, 16 degree-of-freedom (DoF) bipedal robot Tello has the following full-body dynamics:

M(q)q¨+C(q,q˙)q˙+G(q)=Sτ+Jcf,M(q)\ddot{q}+C(q,\dot{q})\dot{q}+G(q)=S^{\top}\tau+J_{c}^{\top}f, (7)

where qq is the robot generalized coordinates vector, MM is the mass matrix, CC is the Coriolis matrix, GG is the gravity vector, SS is the actuated DoF selection matrix, JcJ_{c} is the contact Jacobian matrix, τ\tau is the joint torque vector, and ff is the ground reaction wrench vector.

A popular approximation of Eq. (7) is the single rigid-body model (SRBM) [23, 24]. The SRBM is a simplified model of legged robots with low centroidal inertia isotropy (CII) [22], which neglects the inertial effects introduced by the robot’s limbs. Thus, the torso’s translational and rotational dynamics are only considered. The states of the torso (SRB) are defined as s:=[p,p˙,R,Bω]s:=\left[p,\dot{p},R,^{B}\omega\right]^{\top}, where pp is the CoM position vector, RR the rotation matrix of the body frame relative to the world frame, and ωB{}^{B}\omega is the angular velocity vector expressed in body frame, and {B}\{B\} is the body frame. The SRBM state-space dynamics are the following:

s˙=[p˙p¨R˙ω˙B]=[p˙1mF+agRωBI1B[Rτ~ω^BIBωB]]\dot{s}=\begin{bmatrix}\dot{p}\\ \ddot{p}\\ \dot{R}\\ {}^{B}\dot{\omega}\end{bmatrix}=\begin{bmatrix}\dot{p}\\ \frac{1}{m}F+a_{g}\\ R\cdot{}^{B}\omega\\ {}^{B}I^{-1}[R^{\top}\tilde{\tau}-{}^{B}\hat{\omega}{}^{B}I{}^{B}\omega]\end{bmatrix} (8)

where mm is the robot mass, IB{}^{B}I is the torso inertia matrix in the body frame, aga_{g} is the gravity vector, FF and τ~\tilde{\tau} are translational and rotational components of the net wrench acting on the CoM, respectively, and ()^\hat{\left(\cdot\right)} is the operator that maps an element of 3\mathbb{R}^{3} to the 3D rotation group SO(3)SO(3). The net wrench, \mathcal{F}, arises from the ground reaction forces (GRFs) generated at active contact points, i.e.,

=[Fτ~]=i=1n[𝕀r^i]f~i,\mathcal{F}=\begin{bmatrix}F\\ \tilde{\tau}\end{bmatrix}=\sum_{i=1}^{n}\begin{bmatrix}\mathbb{I}\\ \hat{r}_{i}\end{bmatrix}\tilde{f}_{i}, (9)

where nn is the number of active contact points, 𝕀\mathbb{I} is the identity matrix, rir_{i} represents the vector from the CoM to the ii-th active contact point, while f~i\tilde{f}_{i} denotes the GRF vector at that contact point. For Tello, which has line feet, we assume two contact points at the toe and heel of each foot [24].

III-B Optimization-Based Balance Controller

The balance controller computes the GRF vectors required to reactively track a desired SRBM trajectory. A PD task-space controller is implemented,

d=Kp(qsdqs)+Kd(q˙sdq˙s),\mathcal{F}^{d}=K_{p}(q_{s}^{d}-q_{s})+K_{d}(\dot{q}_{s}^{d}-\dot{q}_{s}), (10)

to compute the desired net wrench, d\mathcal{F}^{d}, to track the desired task-space trajectory, qsdq_{s}^{d} and q˙sd\dot{q}_{s}^{d}. The task-space coordinate vector, qsq_{s}, of the SRB is defined as qs:=[p,Θ]q_{s}:=[p,\Theta]^{\top}, where Θ\Theta is the euler-angle representation of RR. Appropriate coordinate transformations are made to construct qsq_{s} and q˙s\dot{q}_{s} from ss. KpK_{p} and KdK_{d} are diagonal proportional and derivative gain matrices, respectively.

Once d\mathcal{F}^{d} is determined using Eq. (10), a force distribution problem is formulated as a quadratic program (QP). This QP aims to calculate the corresponding GRF vectors at the active contact points, satisfying Eq. (9). Furthermore, the GRF vectors undergo projections from task-space to both the friction cone and the motor-space, where linear inequality actuation constraints are enforced. This accounts for both the parallel actuation scheme of Tello [22] and the effects of friction. GRF vectors are also constrained to not exert a pulling force on the ground. Thus, the actuation-aware force distribution QP takes the following form:

minf~\displaystyle\min_{\tilde{f}} w1d22+w2f~f~022\displaystyle w_{1}\|\mathcal{F}-\mathcal{F}^{d}\|_{2}^{2}+w_{2}\|\tilde{f}-\tilde{f}_{0}\|_{2}^{2} (11)
s.t. |JMJJJTRBWf~|τm(qj,q˙j)(motor lim.)\displaystyle|J_{MJ}^{\top}J_{JT}^{\top}R_{BW}\tilde{f}|\leq\tau_{m}(q_{j},\dot{q}_{j})\quad\text{(motor lim.)}
|f~x,y|μf~z(friction cone)\displaystyle|\tilde{f}_{x,y}|\leq\mu\cdot\tilde{f}_{z}\quad\text{(friction cone)}
f~z0(no-pull)\displaystyle\tilde{f}_{z}\geq 0\quad\text{(no-pull)}

where f~\tilde{f} is the collection of GRF vectors at active contact points, f~0\tilde{f}_{0} is the previous value of f~\tilde{f}, w1w_{1} and w2w_{2} are scalar weighting factors, RBWR_{BW} is a matrix of rotation matrices RR^{\top} that map f~\tilde{f} from the world frame to the body frame, JJTJ_{JT} is a matrix of contact Jacobians that map task-space forces to joint torques, JMJJ_{MJ} is a matrix of topology Jacobians that map joint torques to motor torques, τm\tau_{m} is a vector of configuration and speed-dependent motor torque limits, qjq_{j} and q˙j\dot{q}_{j} are leg joint positions and velocities, respectively, and μ\mu is the friction coefficient. The second term of the cost function in Eq. (11) acts as a solution filter.

Finally, joint torques for legs in stance are set using the SRBM approximation of Eq. (7). However, to maintain the task-space configuration of the robot feet, joint-space posture control is appended to the optimization-based balance controller from Eq. (11). Consequently, the final expression for the joint torques of legs in stance, τst\tau_{st}, is the following:

τst=JJTRBWf~+τpo,\tau_{st}=J_{JT}^{\top}R_{BW}\tilde{f}+\tau_{po}, (12)

where τpo\tau_{po} is vector of posture control joint torques. A sigmoid function is utilized to smoothly transition between stance-leg joint torques and swing-leg joint torques. The computation of τ\tau from Eq. (7) is the summation of smoothed stance-leg and swing-leg joint torques.

III-C Swing-Leg Controller

During SSP, when one of the robot’s legs is in swing, corresponding joint torques cannot be computed using Eq. (11). This is because the leg is no longer in contact with the ground and as a result a GRF cannot be generated. Instead, the goal of the swing-leg controller is to map the robot’s foot position at the start of SSP to a location by the end of SSP that achieves the desired step placement.

This mapping of the robot’s foot during SSP is accomplished via smooth task-space trajectories similar to in [19]. On the transverse plane, specifically the xx or yy axis, the function,

pswx,y(s)=12((1+cos(πs))pswi+(1cos(πs))pswf),p^{x,y}_{sw}(s)=\frac{1}{2}\biggl{(}(1+\cos(\pi s))p^{i}_{sw}+(1-\cos(\pi s))p^{f}_{sw}\biggr{)}, (13)

maps the robot’s initial foot position, pswip^{i}_{sw}, to a final desired foot position, pswfp^{f}_{sw}, along each axis. The phase variable, denoted as s[0,1]s\in[0,1], represents the normalized progression of time during a step. On the other hand, zz-position trajectories are defined differently. The objective is to re-establish contact with the ground at time s=1s=1. The following function,

pswz(s)=pswi+zcl2[1cos(2πs)],p^{z}_{sw}(s)=p^{i}_{sw}+\frac{z_{cl}}{2}\biggl{[}1-\cos(2\pi s)\biggr{]}, (14)

characterized by a maximum step height, zclz_{cl}, attains the intended foot z-position motion.

In this work, swing-leg joint torques, τsw\tau_{sw}, are computed using joint-space PD control. First, utilizing inverse kinematics, Eqs. (13) - (14), along with their derivatives, are transformed into desired joint positions, qjdq^{d}_{j}, and corresponding desired joint velocities, q˙jd\dot{q}^{d}_{j}, trajectories. Then, the joint-space trajectories are tracked using the PD control law,

τsw=Kpsw(qjdqj)+Kdsw(q˙jdq˙j),\tau_{sw}=K^{sw}_{p}(q^{d}_{j}-q_{j})+K^{sw}_{d}(\dot{q}^{d}_{j}-\dot{q}_{j}), (15)

where KpswK^{sw}_{p} and KdswK^{sw}_{d} are diagonal proportional and derivative gain matrices, respectively.

III-D Integration with Whole-Body Dynamic Telelocomotion Framework

The robot controller, composed of the optimization-based balance controller and swing-leg controller, is integrated into our whole-body dynamic telelocomotion framework (Fig. 4). Although the robot controller is capable of operating independently, it was designed to be purely reactive, with locomotion planning being entrusted to the human pilot.

The human locomotion strategy is represented by a 2D LIP trajectory. This trajectory is constructed using dynamic human motion data in the frontal plane, combined with the human walking reference generated from human stepping motion data along the sagittal plane. The trajectory is then scaled to robot proportions. The goal of the robot controller is to reproduce the resulting robot 2D LIP trajectory at the full-body (3D) dynamics level.

The reproduction of the robot’s 2D LIP trajectory is described next. The xx-yy components of the desired robot CoM wrench from Eq. (11) are given by robot forces that track the desired scaled human motion [9]. Additionally, the LIPM assumptions of constant CoM height and the omission of angular dynamics are enforced by Eq. (10). The xx-yy swing-leg trajectory from Eq. (13) is updated to keep dynamic synchronization between the human pilot and the robot at step transitions. Specifically, in the xx-direction, the final desired robot foot position maintains tracking of the scaled HWR at the commencement of the next step. In the yy-direction, the final desired robot foot position maintains equal scaled feet width between the human and robot.

The main challenge in tracking the desired robot 2D LIP trajectory is that the single support time TSSPT_{SSP} is not known a priori. TSSPT_{SSP} is normally fixed in bipedal walking controllers. Thus, to resolve this issue we leverage the consistency of human motion to implement a data-driven adaptive step timing strategy. In particular, we assume human swing-foot z-position motion approximately follows Eq. (14), parameterized by ss (normalized SSP time) and zclz_{cl}. During each step, we employ a damped least-squares method to continuously fit human swing-foot z-position data to Eq. (14). This allows the pilot to control the robot’s step height and vary the robot’s stepping frequency. This ensures the pilot and the robot will touch-down synchronously. The updated estimate of SSP duration is used to update the HWR and the xx-yy swing-leg trajectories for more robust telelocomotion.

IV Simulation Results & Discussion

Refer to caption
Figure 5: Setup of telelocomotion simulation experiments. Real-time telelocomotion of the simulated robot by the human pilot is enabled by UDP communication between our HMI and simulation PC.

To validate our approach to human walking generation and assess the performance of our robot controller in executing the human’s locomotion strategy, we designed simulation telelocomotion experiments. As shown in Fig. 5, during these experiments, our whole-body dynamic telelocomotion framework is implemented in a hardware-simulation set-up. As depicted in Fig. 1, the hardware, our human-machine-interface (HMI), is used by a human pilot to control the dynamics of the bipedal robot Tello in a simulation environment with the MuJoCo physics engine [25].

The first experiment involves velocity tracking, where the human pilot tracks a moving target at increasing speeds (0.1 m/s, 0.2 m/s, and 0.3 m/s) for a total of 6.0 m of synchronized robot walking. A square target is displayed on the ground, and the objective is to generate a walking reference that transitions smoothly between stable walking gaits at different speeds and stepping frequencies. This experiment assesses both the human pilot’s ability to generate a walking reference using Algorithm 1 and the robot controller’s capability to execute the desired human locomotion strategy. The second experiment focuses on synchronized backward walking between the human and robot. It demonstrates similar capabilities as the first experiment but highlights an important telelocomotion behavior for effective navigation in an environment.

In this section, we outline the experimental setup employed in our telelocomotion experiments. The focus is on describing the connection between our HMI and a simulation PC, enabling real-time telelocomotion of a simulated bipedal robot. Subsequently, we showcase the successful completion of experiments 1 & 2 by a trained human pilot, and proceed to analyze the results in the context of the presented methods. Finally, we discuss some limitations of our approach.

IV-A Experimental Setup

During telelocomotion simulation experiments, a trained human pilot utilizes our HMI to dynamically teleoperate the simulated bipedal robot Tello, whose dynamics and corresponding controller are running in the simulation PC in real-time. Communication between the HMI and the simulation PC is established via UDP. As shown in Fig. 5, the human pilot receives visual feedback through a computer monitor, which displays the rendered simulation. This is critical in both experiments as visual feedback is required for velocity target tracking and walking backwards a specific distance.

As illustrated in Fig. 5, our HMI consists of linear actuators, lower-body leg joysticks, a force plate, and a real-time embedded controller. The linear actuators perform dual functions of measuring the human CoM while simultaneously applying haptic forces to the human’s torso. The lower-body leg joysticks capture and measure spatial information pertaining to the locations of the human feet. The force plate measures the net contact wrench applied by the human. The actuators are controlled and the motion capture data is collected by a cRIO-9082 (National Instruments) computer.

The simulation of the full-body dynamics of the bipedal robot Tello, described by Eq. (7), is executed using the physics engine MuJoCo. The simulation is seamlessly integrated into a unified program alongside the Tello control software. This software is parallelized across nine distinct CPUs, with all threads running at a frequency of 1 kHz.

Refer to caption
Figure 6: Results of velocity tracking experiment I. (a) color-coded HWR phase diagram by robot target velocity, (b) robot CoM x-velocity vs. target, (c) human stepping motion, (d) robot CoM x-position vs. target.

IV-B Results & Discussion

Refer to caption
Figure 7: Results of velocity tracking experiment II. Normalized DCM tracking performance in (a) sagittal plane and (b) frontal plane, plus corresponding haptic force feedback to human (c) and (d) normalized CoP tracking during final 10.0 s of the experiment.

Using our framework, depicted in Fig. 4, along with visual feedback, a trained human pilot successfully completed both experiments. These experiments serve as a compelling demonstration, showcasing the synchronized locomotion between a human and a bipedal robot in tasks involving velocity target tracking and backward walking. The completion of these tasks show the effectiveness of our HWR motion generation method and our reactive robot controller.

The results of the velocity target experiment are shown in Figs. 67. Here, the human pilot controls the robot’s locomotion to ensure tracking of the target, as illustrated in Fig. 6(b). Notably, the pilot achieves 6.0 m of synchronized locomotion with the robot, as seen in Fig. 6(d), affirming the viability of sustained telelocomotion using our methods. Importantly, the pilot regulates the robot’s CoM velocity around the target velocity. The key lies in intuitively capturing the human pilot’s locomotion strategy from their stepping motion (Fig. 6(c)). Subsequently, the stepping motion is mapped to a desired walking gait, which serves as the basis for the HWR motion in Fig. 6(a). Indeed, as seen in Fig. 6(a), the abstraction of the HWR as a H-LIPM enables generating dynamically consistent walking references, allowing for smooth transitions between desired walking gaits by considering the step-to-step dynamics of walking. Lastly, the ability of the robot controller to reactively reproduce the desired robot 2D LIP trajectory is highlighted by the normalized DCM tracking performance along the sagittal plane (Fig. 7(a)) and along the frontal plane (Fig. 7(b)). As seen in Fig. 7(c), this leads to consistent and bounded haptic force feedback to the human pilot along the frontal plane. This feedback assists the human pilot in maintaining stepping synchronization with the robot. As discussed in our previous work [15, 9], this is evident from the normalized center of pressure (CoP) synchronization between the two systems, as shown in Fig. 7(d).

Refer to caption
Figure 8: Results of backward walking experiment. (a) color-coded HWR phase diagram by walking type, (b) human swing-foot z-position prediction for a single step, (c) human stepping motion, (d) robot CoM x-position.

The results of the backward walking experiment are shown in Fig. 8. The human pilot dynamically synchronizes their backward stepping with bipedal robot backward walking, covering a distance of 2.0 m (Fig. 8(d)), before proceeding to walk forward to the initial position. Like in the first experiment, the human pilot’s stepping motion is utilized to generate a HWR motion trajectory, as seen in Fig. 8(a) and Fig. 8(c). However, by walking backwards, the pilot commands desired walking gaits characterized by negative orbital energy, as shown in Fig. 8(c). Additionally, to demonstrate the efficacy of our adaptive step time prediction approach, we analyze the first step from the human stepping data. As shown in Fig. 8(b), we continuously monitor the human z-foot position data and fit it to the presumed swing-leg z-trajectory curve. Thus, allowing us to predict both the duration (TSSPT_{SSP}) and the height of the step (zclz_{cl}). The prediction accuracy improves with the accumulation of more data, closely approximating the actual human z-foot position throughout the entire step, particularly towards the end. This accurate prediction ensures the dynamic synchronization between the human and bipedal robot at impact events and gives the human the capability to regulate the frequency of the synchronized stepping between the two bipedal systems.

The obtained results show promise in achieving robust telelocomotion of humanoid robots, enhancing their potential as physical avatars. These results also represent tangible progress from the work in [15], specifically in successfully realizing sustainable synchronization between human and bipedal robot locomotion in a full-body dynamics simulation. Previously, in a reduced-order dynamics simulation, only 1.28 m of forward walking (0.36 m/s max speed) was achieved. However, it is important to acknowledge the limitations inherent to this work. First, the walking speed of the robot is currently still limited to approximately 0.35 m/s, which, when considering a robot like Tello, corresponds to a human walking speed of 0.5 m/s. While it is possible for the robot to reach faster walking speeds, it becomes challenging for the pilot to sustain telelocomotion at these speeds. The most common failure is the robot lagging behind the HWR due to prolonged periods of DSP by the human that increases the error in normalized DCM between the two systems at the commencement of the next step. Second, it is essential to emphasize that significant training is required for a pilot to reach the demonstrated level of performance. For instance, in the velocity target tracking experiment, the pilot faced difficulties in abruptly stopping the robot from a walking speed of 0.3 m/s. Similarly, transitioning from backward walking to forward walking in the second experiment proved equally challenging. Lastly, it is worth noting that the telelocomotion behaviors enabled by our framework are presently constrained due to the utilization of linear pendulum models to abstract the locomotion of both the human and bipedal robot. To attain a broader spectrum of behaviors, including loco-manipulation tasks, it will be necessary to employ more dynamically rich reduced-order models that can fully exploit the human’s whole-body dynamics strategy.

V Conclusion

This paper presents an approach that enhances dynamically consistent walking trajectories within a comprehensive whole-body dynamic telelocomotion framework. By integrating the hybrid and underactuated characteristics of walking, we effectively capture the locomotion strategy of a human pilot through their stepping behavior. A reactive robot controller enables telelocomotion of the full-body dynamics of a bipedal robot. Real-time telelocomotion simulation experiments demonstrate the synchronization between a trained human pilot and a bipedal robot, including tracking a moving target and synchronized backward walking.

Future work will focus on implementing these methods on the real humanoid robot Tello, enabling dynamic control for navigating 3D environments. We will develop a human motion mapping for robot turning and address performance limitations related to walking speed. Additionally, we aim to incorporate human predictive intent into our human walking interface, leveraging human intelligence to a greater extent.

References

  • [1] Kourosh Darvish, Luigi Penco, Joao Ramos, Rafael Cisneros, Jerry Pratt, Eiichi Yoshida, Serena Ivaldi, and Daniele Pucci. Teleoperation of humanoid robots: A survey. IEEE Transactions on Robotics, 2023.
  • [2] Luigi Penco, Nicola Scianca, Valerio Modugno, Leonardo Lanari, Giuseppe Oriolo, and Serena Ivaldi. A multimode teleoperation framework for humanoid loco-manipulation: An application for the icub robot. IEEE Robotics & Automation Magazine, 26(4):73–82, 2019.
  • [3] Yasuhiro Ishiguro, Kunio Kojima, Fumihito Sugai, Shunichi Nozawa, Yohei Kakiuchi, Kei Okada, and Masayuki Inaba. High speed whole body dynamic motion experiment with real time master-slave humanoid robot system. In 2018 IEEE International Conference on Robotics and Automation (ICRA), pages 5835–5841, 2018.
  • [4] Junheng Li and Quan Nguyen. Kinodynamics-based pose optimization for humanoid loco-manipulation. arXiv preprint arXiv:2303.04985, 2023.
  • [5] Alphonsus Adu-Bredu, Grant Gibson, and Jessy W Grizzle. Exploring kinodynamic fabrics for reactive whole-body control of underactuated humanoid robots. arXiv preprint arXiv:2303.04279, 2023.
  • [6] Joao Ramos and Sangbae Kim. Dynamic locomotion synchronization of bipedal robot and human operator via bilateral feedback teleoperation. Science Robotics, 4(35):eaav4282, 2019.
  • [7] Mohamed Elobaid, Yue Hu, Giulio Romualdi, Stefano Dafarra, Jan Babic, and Daniele Pucci. Telexistence and teleoperation for walking humanoid robots. In Intelligent Systems and Applications: Proceedings of the 2019 Intelligent Systems Conference (IntelliSys) Volume 2, pages 1106–1121. Springer, 2020.
  • [8] Joris Verhagen, Xiaobin Xiong, Aaron D Ames, and Ajay Seth. From human walking to bipedal robot locomotion: Reflex inspired compensation on planned and unplanned downsteps. In 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 01–08. IEEE, 2022.
  • [9] Joao Ramos and Sangbae Kim. Humanoid dynamic synchronization through whole-body bilateral feedback teleoperation. IEEE Transactions on Robotics, 34(4):953–965, 2018.
  • [10] Yasuhiro Ishiguro, Tasuku Makabe, Yuya Nagamatsu, Yuta Kojio, Kunio Kojima, Fumihito Sugai, Yohei Kakiuchi, Kei Okada, and Masayuki Inaba. Bilateral humanoid teleoperation system using whole-body exoskeleton cockpit tablis. IEEE Robotics and Automation Letters, 5(4):6419–6426, 2020.
  • [11] Luka Peternel and Jan Babič. Learning of compliant human–robot interaction using full-body haptic interface. Advanced Robotics, 27(13):1003–1012, 2013.
  • [12] Stefano Dafarra, Kourosh Darvish, Riccardo Grieco, Gianluca Milani, Ugo Pattacini, Lorenzo Rapetti, Giulio Romualdi, Mattia Salvi, Alessandro Scalzo, Ines Sorrentino, et al. icub3 avatar system. arXiv preprint arXiv:2203.06972, 2022.
  • [13] Max Schwarz, Christian Lenz, Andre Rochow, Michael Schreiber, and Sven Behnke. Nimbro avatar: Interactive immersive telepresence with force-feedback telemanipulation. In 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 5312–5319. IEEE, 2021.
  • [14] Sunyu Wang and Joao Ramos. Dynamic locomotion teleoperation of a reduced model of a wheeled humanoid robot using a whole-body human-machine interface. IEEE Robotics and Automation Letters, 7(2):1872–1879, 2021.
  • [15] Guillermo Colin, Youngwoo Sim, and Joao Ramos. Bipedal robot walking control using human whole-body dynamic telelocomotion. In 2023 IEEE International Conference on Robotics and Automation (ICRA), pages 12191–12197, 2023.
  • [16] Yachen Zhang and Ryo Kikuuwe. Com shifter and body rotator for step-by-step teleoperation of bipedal robots. IEEE Access, 11:25786–25800, 2023.
  • [17] Ata Otaran and Ildar Farkhatdinov. Walking-in-place foot interface for locomotion control and telepresence of humanoid robots. In 2020 IEEE-RAS 20th International Conference on Humanoid Robots (Humanoids), pages 453–458. IEEE, 2021.
  • [18] Jessy W Grizzle, Christine Chevallereau, Ryan W Sinnet, and Aaron D Ames. Models, feedback control, and open problems of 3d bipedal robotic walking. Automatica, 50(8):1955–1988, 2014.
  • [19] Yukai Gong and Jessy Grizzle. Angular momentum about the contact point for control of bipedal locomotion: Validation in a lip-based controller. arXiv preprint arXiv:2008.10763, 2020.
  • [20] Xiaobin Xiong and Aaron Ames. 3-d underactuated bipedal walking via h-lip based gait synthesis and stepping stabilization. IEEE Transactions on Robotics, 38(4):2405–2425, 2022.
  • [21] Xiaobin Xiong and Aaron D Ames. Orbit characterization, stabilization and composition on 3d underactuated bipedal walking via hybrid passive linear inverted pendulum model. In 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 4644–4651. IEEE, 2019.
  • [22] Youngwoo Sim and Joao Ramos. Tello leg: The study of design principles and metrics for dynamic humanoid robots. IEEE Robotics and Automation Letters, 7(4):9318–9325, 2022.
  • [23] Jared Di Carlo, Patrick M Wensing, Benjamin Katz, Gerardo Bledt, and Sangbae Kim. Dynamic locomotion in the mit cheetah 3 through convex model-predictive control. In 2018 IEEE/RSJ international conference on intelligent robots and systems (IROS), pages 1–9. IEEE, 2018.
  • [24] Junheng Li and Quan Nguyen. Force-and-moment-based model predictive control for achieving highly dynamic locomotion on bipedal robots. In 2021 60th IEEE Conference on Decision and Control (CDC), pages 1024–1030. IEEE, 2021.
  • [25] Emanuel Todorov, Tom Erez, and Yuval Tassa. Mujoco: A physics engine for model-based control. In 2012 IEEE/RSJ international conference on intelligent robots and systems, pages 5026–5033. IEEE, 2012.