Human Tracking - first tips

There are some different systems for estimating the pose of a human in a real time video (for /human_tracker node). The most known (and better documented) is OpenPose.

OpenPose allows:

  • 2D real-time multi-person keypoint detection
  • 3D real-time single-person keypoint detection (compatible with PointGrey cameras, requiring multiple views)

The best advantage of OpenPose is that it can also detect the hand shape, being useful for the /human_decision node as well. With many keypoints all over the human body, including its hand, it is easily extracted the human decision based upon the hand gestures.

link to their github: OpenPose github

link to their paper: OpenPose paper

Other alternatives for this human tracking purpose are HRNet and MMPose, both of them based on PyTorch (a python library that eases building deep learning projects). These two algorithm had used COCO and MPII human poses datasets for training their trackers.

The HRnet has proved to be more accurate than OpenPose, however it do not take the human hand in consideration, and it is not so well integrated in different hardware and sources as it is OpenPose. For a more detailed comparison, see this.

The real challenge, as far as I understand, it will be how to get the 3D spacial pose of the human body, from this 2D keypoints of one camera image... More to be discussed further!



Comments

Popular posts from this blog

Remote Control of UR10e via MoveIt (ROS)

Installing External Control URCap on robot Teach Pendant

UR10e control architecture