Skip to content

Latest commit

 

History

History
110 lines (77 loc) · 4.92 KB

File metadata and controls

110 lines (77 loc) · 4.92 KB

Tensegrity Robot Proprioceptive State Estimation with Geometric Constraints

Wenzhe Tong1    Tzu-Yuan Lin1    Jonathan Mi1    Yicheng Jiang1    Maani Ghaffari1    Xiaonan Huang1
1University of Michigan, Ann Arbor   

About

tensegrity robot state estimator is a novel framework designed to estimate the 3-bar tensegrity robot shape and pose.

  • To the best of our knowledge, this work is the first proprioceptive Invariant Extended Kalman Filter (InEKF) state estimator that can estimate both shape and pose of the 3-bar tensegrity robot.
  • We introduced a measurement model tailored to the kinematics of tensegrity robots.
  • We incorporated geometric properties of the 3-bar tensegrity robot as constratins in the robot shape reconstruction process using constrained optimization.

Framework

Both real-world or simulated IMU and cable length sensors are inputs into an optimization-based robot shape reconstruction algorithm. The reconstructed shape provides the positions of the robot’s endcaps in the body frame. Next, the computed kinematics, based on the contact points between the endcaps and ground, are utilized within a contact-aided Invariant EKF to estimate the robot pose. Finally, the global endcap positions are computed by transforming the reconstructed shape into the body pose within a global frame.

Updates

  • [2024.10] - Our paper is available on arXiv.

Outline

TODO List

  • Initial README
  • ArXiv paper release
  • Initial code release

Citation

If you find this work helpful for your research, please kindly consider citing our papers:

 @article{tong2024tensegrity,
            title={Tensegrity Robot Proprioceptive State Estimation with Geometric Constraints},
            author={Tong, Wenzhe and Lin, Tzu-Yuan and Mi, Jonathan and Jiang, Yicheng and Maani Ghaffari and Huang, Xiaonan},
            journal={arXiv preprint arXiv:2410.24226},
            year={2024}
          }

License

This work is under the MIT License, while some specific implementations in this codebase might be with other licenses. Kindly refer to LICENSE for a more careful check, if you are using our code for commercial matters.