JCSE, vol. 18, no. 1, pp.47-56, 2024
DOI: http://dx.doi.org/10.5626/JCSE.2024.18.1.47
Real-Time Retargeting of Human Poses from Monocular Images
and Videos to the NAO Robot
Oscar Burga, Jonathan Villegas, and Willy Ugarte
Department of Computer Science, Universidad Peruana de Ciencias Aplicadas, Lima, Peru
Abstract: To this point, there has been extensive research investigating human-robot motion retargeting, but the vast majority of existing methods rely on sensors or multiple cameras to detect human poses and movements, while many other methods are not suitable for usage on real-time scenarios. The current paper presents an integrated solution for performing realtime human-to-robot pose retargeting utilizing only regular monocular images and video as input data. We use deep learning models to perform three-dimensional human pose estimation on the monocular images and video, after which we calculate a set of joint angles that the robot must utilize to reproduce the detected human pose as accurately as possible. We evaluate our solution on Softbank?셲 NAO robot and show that it is possible to reproduce promising approximations and imitations of human motions and poses on the NAO robot, although it is subject to the limitations imposed by the robot?셲 degrees of freedom, joint constraints, and movement speed limitations.
Keyword:
Human pose estimation; Humanoid robot; Motion retargeting; Geometry; Vectors
Full Paper: 81 Downloads, 320 View
|