This is a submodule of the A2Perf project. For complete documentation and usage instructions, please refer to the main A2Perf README.
Further development (new features, bug fixes etc) happen in the master branch. The 'paper' branch of this repository contains the original code accompanying the paper:
"Learning Agile Robotic Locomotion Skills by Imitating Animals",
by Xue Bin Peng et al. It provides a Gym environment for training a simulated quadruped robot to imitate various reference motions, and example training code for learning the policies.
Project page: https://xbpeng.github.io/projects/Robotic_Imitation/index.html
We use this repository with Python 3.7 or Python 3.8 on Ubuntu, MacOS and Windows.
- Install MPC extension (Optional)
python3 setup.py install --user
Install dependencies:
- Install MPI:
sudo apt install libopenmpi-dev
- Install requirements:
pip3 install -r requirements.txt
and it should be good to go.
Build Docker
TBD
To train a policy, run the following command:
python3 motion_imitation/run.py --mode train --motion_file motion_imitation/data/motions/dog_pace.txt --int_save_freq 10000000 --visualize
--mode
can be eithertrain
ortest
.--motion_file
specifies the reference motion that the robot is to imitate.motion_imitation/data/motions/
contains different reference motion clips.--int_save_freq
specifies the frequency for saving intermediate policies every n policy steps.--visualize
enables visualization, and rendering can be disabled by removing the flag.- the trained model and logs will be written to
output/
.
For parallel training with MPI run:
mpiexec -n 8 python3 motion_imitation/run.py --mode train --motion_file motion_imitation/data/motions/dog_pace.txt --int_save_freq 10000000
-n
is the number of parallel.
To test a trained model, run the following command
python3 motion_imitation/run.py --mode test --motion_file motion_imitation/data/motions/dog_pace.txt --model_file motion_imitation/data/policies/dog_pace.zip --visualize
--model_file
specifies the.zip
file that contains the trained model. Pretrained models are available inmotion_imitation/data/policies/
.
motion_imitation/data/motions/
contains different reference motion clips.motion_imitation/data/policies/
contains pretrained models for the different reference motions.
For more information on the reference motion data format, see the DeepMimic documentation
To start, just clone the codebase, and install the dependencies using
pip install -r requirements.txt
Then, you can explore the environments by running:
python3 -m motion_imitation.examples.test_env_gui --robot_type=A1 --motor_control_mode=Position --on_rack=True
The three commandline flags are:
robot_type
: choose between A1
and Laikago
for different robot.
motor_control_mode
: choose between Position
,Torque
for different motor control modes.
on_rack
: whether to fix the robot's base on a rack. Setting on_rack=True
is handy for debugging visualizing open-loop gaits.
Additionally, the codebase can be directly installed as a pip package. Just run:
pip3 install motion_imitation --user
Then, you can directly invoke the default gym environment in Python:
import gymnasium as gym
env = gym.make('motion_imitation:A1GymEnv-v1')
Note that the pybullet rendering is slightly different from Mujoco. To enable GUI rendering and visualize the training process, you can call:
import gymnasium as gym
env = gym.make('motion_imitation:A1GymEnv-v1', render=True)
which will pop up the standard pybullet renderer.
And you can always call env.render(mode='rgb_array') to generate frames.
Since the SDK from Unitree is implemented in C++, we find the optimal way of robot interfacing to be via C++-python interface using pybind11.
To start, build the python interface by running the following:
cd third_party/unitree_legged_sdk
mkdir build
cd build
cmake ..
make
Then copy the built robot_interface.XXX.so
file to the main directory (where you can see this README.md file).
Since the Unitree SDK requires memory locking and high-priority process, which is not usually granted without sudo, add the following lines to /etc/security/limits.conf
:
<username> soft memlock unlimited
<username> hard memlock unlimited
<username> soft nice eip
<username> hard nice eip
You may need to reboot the computer for the above changes to get into effect.
Test the python interfacing by running: 'sudo python3 -m motion_imitation.examples.test_robot_interface'
If the previous steps were completed correctly, the script should finish without throwing any errors.
Note that this code does not do anything on the actual robot.
To see the whole-body MPC controller in sim, run:
python3 -m motion_imitation.examples.whole_body_controller_example
To see the whole-body MPC controller on the real robot, run:
sudo python3 -m motion_imitation.examples.whole_body_controller_robot_example
This repo was developed at Google Robotics and is maintained by one of its members, Erwin Coumans. The original Motion Imitation code was written by Jason Peng as part of an internship and student researcher at Google Robotics. Some MPC parts for A1 and running on real A1 are written by Yuxiang Yang, a former resident researcher at Google Robotics.
Disclaimer: This is not an official Google product.