/ 

Compliant Biped Humanoid Robots

- Contents -

What's New

Research Objectives

Platform

Framework

Results:

References

Contributors

Acknowledgements


What's New:

  • (2010/9/8) This page has been moved from a ATR server to Ritsumeikan server to make it easy for the administrator to update the contents. All the contents (photo, videos) are with ATR and/or NICT.
  • (2010/5/5) Our work "From compliant balancing to dynamic walking on humanoid robot: Integration of CNS and CPG" was selected as the Finalist for the IEEE ICRA 2010 Best Conference Video Award.
  • (2010/5/5)
    This work has been introduced in 2010 IEEE International Conference on Robotics and Automation (ICRA) Workshop on active force control for humanoid robots.
    We invited top runners from BostonDynamics, Honda, Disney Research, ATR, DLR, USC, Stanford Univ. to discuss together on state-of-the-art force control technologies applied to humanoid/legged robots.
  • (2009/5/16) Our recent paper "Integration of multi-level postural balancing on humanoid robots" was selected as the Finalist for the IEEE ICRA 2009 Best Conference Paper Award.
  • (2008/9/10) Young Investigation Excellence Awards given by Robotics Society of Japan.
  • (2008/5/23) A new control/learning framework presented at IEEE ICRA2008.
  • (2008/5/4) A baseball batting demo on CB-i introduced in NHK Special.
  • (2007/12/6) Article "Balancing Robot Can Take a Kicking" on Slashdot
  • (2007/12/6) Article "Flexible-jointed robot is no pushover" on NewScientistTech
  • (2007/10) The full-paper on compliant balancing and interaction under unknown disturbances was published on IEEE TRO, vol.23, no.5.
  • (2006/12/4) Some experimental results were presented at IEEE-RAS Humanoids2006.
  • (2006/5/27) Full-body compliant balancing under unknown disturbance was achieved on a SARCOS biped humanoid robot.


Research Objectives:

  • Understanding human motor control and learning especially focusing on postural control and biped walking
  • Realization of hyper performance humanoid/legged robots
  • Elucidation of versatile locomotion with dynamical stability
  • Development of computationally-plausible assistive devices for humans


Platform:

dof
system

For more technical details of the platform, see:

Cheng, G., Hyon, S., Morimoto, J., Ude, A., Hale, J. G., Colvin, G., Scroggin, W. and Jacobsen, S. C.,
CB: A humanoid research platform for exploring neuroscience,
Advanced Robotics, vol.21, no.10, pp.1097-1114, 2007.


Framework:

We test our hypothesis of motor control and learning of humans by testing possible hierarchical computation and body structures on a full-body humanoid robots, especially focusing on the function of the task-space controller, joint-space controller, and musculoskeletal systems, as shown in the figure below.
The former is supposed to be in the central nervous system (CNS), and the second in the central pattern generator (CPG) or lower reflex centers.
Not only testing hypothesis of each module, but also investigating how these three modules are combined effectively to achieve optimal motor control performance in different contexts is our primary research interest.

framework


Results:


Full-body torque control

We first implemented joint torque controllers using the force sensors attached to the joints.
This enabled us to develop several motor control algorithms in which the control input is defined as the joint torque.
Also this enabled us to functionally emulate human musculo-skeletal system, which can control load and stiffness simultaneously.

The extra benefit is that we can easily simulate the robot behaviors on multi-body dynamic simulators without considering the actuator dynamics which are required in conventional position-control-based robots.
The joint torque controller is designed according to the actuator dynamics including hydrostatic dynamics.


With the full-body torque controller, the robot can make its limbs free, or gravity-compensated to passively follow external forces applied by humans.
This is a fundamental function for natural and safety interaction between humans and robots.


Moreover, with the gravity compensation, fast joint motions can be achieved by superposing joint trajectory tracking controllers with very roughly-tuned PD gains.

Since the gravity compensation requires internal model of the body and vestibular feedback, it is natural to assume the function is implemented in the central nervous systems (CNS) of humans, not in any other low-level nervous systems.


Full-body force control / Gravity adaptation


Humans can stably stand on rough terrain by applying forces to each foot or hand.
So, we developed a practical control algorithm that can optimally control the multiple contact forces between the robot and the ground simultaneously.

Our computational procedure is simple as follows:
1) Compute the desired global ground applied force (GAF),
2) Optimally distribute that force into the multiple contact points,
3) Convert the multiple contact forces into the whole-body joint torques,
4) Add joint-wise damping to supress internal motions.

The step 4) is called as a passivity-based redundancy resolution, which is a simple, but important concept in redundant robots (but tuning appropriate damping is not straightforward).


The full-body gravity compensation while standing is achieved by commanding the anti-gravity force as the desired global GAF in the step 1).
Then, the robot behaves like "clay"; passively adapts to external forces and surrounding surface (see the next section).

The left figure shows the robot is passively following to the external forces while standing by both feet.
This is actually used for teaching motions from human to the robot (see below).


Full-body compliant balancing


We can specify some desired center of mass (CoM) motions (time course of acceleration, velocity, position), and convert them into the desired GAF in a feedforward or feedback form in the algorithm 1) in the last section. If we set the desired CoM position to zero, the algorithm serves as a full-body balancing controller.


Since the robot has many redundant joints, the balancing behavior is compliant in the sence that the robot can move freely in the range of motions while keeping the CoM to the desired position, as can be seen in the left figures.

If we superpose some joint-trajectory tracking controller to the null-space of the task-space controller, the desired joint motions are achieved with the secondary priority.
Practically, simple superposition without null-space projection is useful to some extent.

[Simulation]
  • Torso bending (0.5 Hz) showing the superposed torso motion is automatically compensated by the balancing controller.

    The load distribution to the both feet is optimally determined by the algorithm 2).
    When the robot CoM is shifted to one side, the center of pressure (CoP) is shifted to the same side, and the contact forces distribution is shifted accordingly.This may cause the contact foces of the other side foot to zero, resulting in the lift off when the leg is fully extended as can be seen in the left figure, where the human is pulling the robot to left side.
    After the touchdown, the robot is balancing at the new contact points.


    Thanks to the vestibular feedback, the robot can balance on randomly moving seesaw.
    The vestibular feedback is used in the algorithm 4) and the CoM estimation.
    When the seesaw changes its inclincation, the robot foot instantenously lose its contact if the robot is stiff.
    This is not our case because our controller is force-based, implemented the optimal load distribution, hence has natural terrain adaptability (see also the next section).


    We can specify some CoM trajectorys, for example, sinusoidal ones.
    Periodic sway, or squat motions can be easily created.
    The left figure shows a circular CoM motion in XY plane.

    [Simulations]
  • Squat with the upper-body swinging showing the superposition of the CoM tracking controller and the (joint-level) trajectory tracking controller of the torso
  • High-speed squat on one-foot showing the superposition of the CoM tracking controller and the (task-space) position controller of the one foot

    Also, the desired contact points can be added or removed so that the robot can take toe-off or heel-off contact.[Simulations]
  • Balancing with variable contact
  • Diagonal sway

    Some experimental videos are available from the suplemental material for our papers:
  • IEEE TRO, vol.23, no.5
  • IEEE TRO, vol.25, no.1
    (If it doesn't work, see video_08-0033.mp4 (11 MB) with the instruction readme.txt ).


Phase transition / Static walking


The robot can transit from double support (DS) to single support (SS) by shifting the desired CoM position and making the contact switching.
For example, we can use a smooth target CoM trajectory (sinusoidal).
When the CoM velocity exceeds the threshold determined by the saddle condition of the inverted pendulum model, we remove the contact from one foot.

The transition from SS to DS (touchdown) is done in a feedforward manner; the robot simply switches its controller from SS balancing to DS balancing, regardless of the actual touchdown, once the decision of the phase transition has been made (hence the prediction is important).
The swinging leg trajectory is not given at all.
This is a powerful strategy that compliant robots only can take.
Thanks to the innate terrain adaptability of our controller, the robot can land or even stand on wooden blocks of the moderate size.


If we do not specify the CoM height or the knee angle, the supporting leg has compliance even during the one-foot balancing because the robot is gravity compensated.
In the same way, the swinging leg can bee freely moved by hand unless some trajectry tracking controllers are applied to it.


The ability of balancing at single and double support phases and of transition between them means that the robot can perform static, or quasi-dynamic biped walking.
Actually, forward/backward walking have been achieved as shown in the left figure without any technical difficulties.
Wherein, the swinging foot position is commanded by the operator's key inputs, and the desired CoM position is fixed to the center of the supporting foot, except for when the transition from SS to DS is made (therefore this is a quasi-dynamic walking).


With exactly the same controller, the robot can also climb steps with the moderate height (kicking motion is required for higher steps).


Push-recovery control


Physiologists have been discussing human push-recovery strategies based on experimental data of humans.
The left figure shows the three main strategies.
There are ankle strategy, hip strategy, and step strategy.
Naturally, it is of our interest how and where these strategies are implemented in our nervous system and body, and when these are switched according to various unstable situations.
Therefore, we studied such strategies from control point of view and experimentally implemented them on our humanoid robot.

The ankle strategy can be extended to the full-body balancing controller decribed above because our controller can be considered as the full-body reflex including the reflex by the ankle stiffness, where we used static internal model and simple dynamic inverted pendulum model.

On the other hand, the hip strategy is useful when the CoM is out of the support region.
In this case, the system becomes underactuated and relatively high nervous cost may be required.
We implemented the hip strategy based on an underactuated double-pendulum model.

We didn't make any switching between the ankle and hip strategies, but combined the both with the simple torque superposition.


The left figure shows a strong push-recovery motion using this simple integration of the hip strategy and ankle strategy.
One can see the rotational motion of the upper body is relatively larger than when only the "ankle strategy" is employed.
With this integration, the region of attraction of CoM was enlarged (but the computational cost increased).

[Simulations]
  • Integrated balancer showing the combined ankle and hip strategies

    Furthermore, we implemented a heuristic step strategy, where the robot takes step to recover the balance at the next support phase when the push is too strong (threshold has been set).Herein, the one leg is swung symmetric to the CoM, hence the motion of the swinging leg is synchronized to that of the CoM.
    The robot may take multiple steps according to the magnitude of the push, which is shown in the simulation movies.[Simulations]
  • Disturbance rejection with stepping The experimental videos are available from the suplemental material for our papers:
    IEEE ICRA, 2009


Dynamic biped walking


Up to simulations, it is found that biped walking that is robust to the external disturbances can be simply achieved by superposing the above mentioned symmetric step strategy onto the full-body balancing controller under the velocity control mode.
If we turn off the balancing controller, then the robot is purely gravity compensated.
With the symmetric stepping, however, the robot can walk at almost constant velocity.
When pushed, the robot changes its moving direction, just like a "ball" rolling on a flat surface (globally stable).
We call this Symmetric Walking Control.
The theoretical discussion can be found in our paper.

When we activate the velocity controller together with the above controller, then the robot can walk at the specified speed and direction.
This is our passivity-based control solution for globally asymptotically stable walking (gait generation).

[Simulations]
  • Symmetric Walking Control: Externally-driven walking showing the walking gait is (Lyapnov) stable
  • Symmetric Walking Control: Speed-regulated walking showing the walking gait is asymptotically stable
    Although the above simulations show the effectiveness of our control method for stable walking, the controller requires relatively precise tracking performance of both swinging leg and CoM, whereas task-space tracking controllers, in practice, usually suffer from modeling/sensing errors.
    As a practical solution, we proposed to combine joint-space controllers as shown in the above control framework.
    Specifically, we encode the self-generated quasi-dynamic walking motions into some joint-, or higher-level oscillators.
    Then, we adapt the oscillator model through learning to increase the speed of the walking phase, leading to dynamic CNS+CPG walking.
    For the learning from static to dynamic motions, see the next section.
    The left figure shows our preliminary reslt.


Learning dynamic full-body motions from static execution


The left figure shows the new learning framework shown in our paper. This is composed of
three modules: C1 task-space control center, C2 joint pattern
generator, C3 supervised adaptive/learning center
.

C1 computes the necessary joint torques via Jacobian
transpose, which are required to follow some position of
force trajectories given in the task-space (desired trajectory).
This requires forward kinematics based on internal
or external sensory feedback, indicated by FB1 and FB2 in
the figure. As being static, this controller has difficulty in
executing dynamic tasks when used alone.

C2 is arranged to the joint space and learns the joint
trajectories while the robot is slowly executing tasks. The
learned joint trajectories are called "reference trajectories".
C2 also generates attractive force field which pulls
the joint trajectories to the reference trajectories by specified
joint stiffness.

C3 learns feedforward torque to compensate the dynamics,
which cannot be treated by C1, according to the task-space
tracking error.
C3 also adaptively tunes the joint stiffness according to the
task-space tracking errors as the dynamics learning proceeds.


As an example, we simulated online learning of fast squatting motions,
which are iteratively acquired from the static execution.

[Simulations]


Teaching full-body motions from human to robot




Manipulation


[Simulations] 1) Task-space tracking of the hand position/orientation


Baseball batting

For the background and experimental video, see the website of the ICORP Computational Brain Project.


We first studied a simple online learnig algorithm with sequential least square method to predict the meeting time t^* and the position of the ball (x^*, z^*) on the cross section (see the left figure).

The 3D ball position is recognized by the robot eye cameras (two foveal cameras), and the throwing timing is detected by the incoming velocity.
During the ball detection and prediction the robot is supposed to be in quiet stance.


Fast swinging motions are generated as follows.
1) Apply cheap balancing control combined with G-comp and stiffness,
2) Superpose a simple PD-based tracking control for a family of minimum jerk upper-body trajectories whose via-points are set acc. the initial posture and different meeting positions of the bat.

[Simulations]
  • Fast batting with balancing
    The left picture shows that the robot is successfully hitting the ball.
    The project is ongoing to improve the ball prediction and the performance of hitting.


References


Technical Contributors

  • Force controller implementation was done with Dr. Gordon Cheng. He provided the realtime communication softwares for both robots.
  • Idea of frequency modulation of CPG was borrowed from Dr. Jun Morimoto. He first demonstrated a biped walking experiment using the idea.
    He also provided a template of the simulator (SD/FAST + OpenGL).
  • We maintain our hardwares with the help of Mr. Nao Nakano.
  • Vision part of the initial baseball demo was done with Dr. Jan Moren.

Acknowledgements

  • All the works here are supervised by Dr. Mitsuo Kawato, the Director of Computational Neuroscience Laboratories, ATR, Japan.
  • The robots are with ATR or NICT. All the works here are conducted by S. Hyon at ATR or NICT as a researcher (2005-2009), or as a visiting researcher (2010-).
  • We thank the support of National Institute of Information and Communications Technology (NICT). The humanoid robot I-1 (blue one) is developed by NICT in the collaborative research project with ATR.
  • We thank the support of Japan Science and Technology Agency (JST). The humanoid robot CB-i (gray one) is developed by JST, ICORP, Computational Brain Project (2004.1-2009.3).
  • We wish to acknowledge the support of the members of SARCOS. The both robots were fablicated by SARCOS. Dr. Gordon Cheng organized this development as the department head of HRCN, CNS, ATR.
  • We wish to acknowledge Sankyoku Co., Ltd. and Mori Kougyou Co., Ltd. for their continous support on hydraulic systems for more than decade.

2020/06/11

PAGE TOP