Learning-assisted Multi-IMU Proprioceptive State Estimation for Quadruped Rob

Read the full article See related articles

Listed in

This article is not in any list yet, why not save it to one of your lists.
Log in to save this article

Abstract

This paper presents a learning-assisted approach for state estimation of quadruped robots using observations of proprioceptive sensors including multiple inertial measurement units (IMUs). Specifically, one body IMU and four additional IMUs attached to each calf link of the robot are used for sensing the dynamics of the body and legs, in addition to joint encoders. The extended Kalman filter (EKF) is employed to fuse sensor data to estimate the robot’s states in the world frame and enhance the convergence of EKF. To circumvent the requirements for the measurements from the motion capture (mocap) system or other vision systems, the right-invariant EKF (RI-EKF) is extended to employ the foot IMU measurements for enhanced state estimation, and a learning-based approach is presented to estimate the vision system measurements for the EKF. 1D convolutional neural networks (CNN) are leveraged to estimate required measurements using only the available proprioception data. Experiments on the real data of a quadruped robot demonstrate that proprioception can be sufficient for state estimation, and the proposed learning-assisted approach without using data from vision systems can achieve competitive accuracy compared to EKF using the mocap measurements and smaller estimation errors than RI-EKF using multi-IMU measurements.

Article activity feed