Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2021 Aug 5;21(16):5301.
doi: 10.3390/s21165301.

Object Manipulation with an Anthropomorphic Robotic Hand via Deep Reinforcement Learning with a Synergy Space of Natural Hand Poses

Affiliations

Object Manipulation with an Anthropomorphic Robotic Hand via Deep Reinforcement Learning with a Synergy Space of Natural Hand Poses

Patricio Rivera et al. Sensors (Basel). .

Abstract

Anthropomorphic robotic hands are designed to attain dexterous movements and flexibility much like human hands. Achieving human-like object manipulation remains a challenge especially due to the control complexity of the anthropomorphic robotic hand with a high degree of freedom. In this work, we propose a deep reinforcement learning (DRL) to train a policy using a synergy space for generating natural grasping and relocation of variously shaped objects using an anthropomorphic robotic hand. A synergy space is created using a continuous normalizing flow network with point clouds of haptic areas, representing natural hand poses obtained from human grasping demonstrations. The DRL policy accesses the synergistic representation and derives natural hand poses through a deep regressor for object grasping and relocation tasks. Our proposed synergy-based DRL achieves an average success rate of 88.38% for the object manipulation tasks, while the standard DRL without synergy space only achieves 50.66%. Qualitative results show the proposed synergy-based DRL policy produces human-like finger placements over the surface of each object including apple, banana, flashlight, camera, lightbulb, and hammer.

Keywords: anthropomorphic robotic hand; deep reinforcement learning; natural hand poses; object grasping; object relocation; synergy space.

PubMed Disclaimer

Conflict of interest statement

The authors declare no conflict of interest.

Figures

Figure 1
Figure 1
The proposed synergy–based DRL for natural object manipulation with a synergy space of object haptic information and a deep regressor to derive natural hand poses of the ADROIT anthropomorphic robotic hand.
Figure 2
Figure 2
From human demonstrations (top row), haptic information is shown in point clouds (second row). CNF encodes the haptic information, qϕ into a synergy space representation, 𝓏. The deep regressor, Rθ𝓏 estimates natural hand poses (final row) in time series.
Figure 3
Figure 3
Reconstruction of object haptic information using the synergy-based encoder-decoder via CNF for three synergy dimensions. The illustrations are derived from training iterations: (a), (c), and (e) point clouds shown in gray are the ground truth haptic maps, the (b), (d), and (f) point clouds in blue are the reconstructed maps from CNF models with dimensions of z32,z16,z4, respectively.
Figure 4
Figure 4
The mean squared error (c) between the joint angle distributions of the haptic map dataset, (a) and the angles distributions estimated by the regressor (b) for each finger in the anthropomorphic hand. A lower error indicates higher confidence in estimating the joint angles of a natural grasping pose.
Figure 5
Figure 5
The left column shows the average sum of rewards from training the manipulation task with the proposed synergy-based DRL in purple and synergy-less standard DRL in blue. The solid line represents the mean and the shadow area of the standard deviation of the rewards. The right column shows the time-series frames from grasping and relocation of each object from apple, banana, hammer, lightbulb, flashlight, and camera with the standard DRL in the blue frames of (a,c,e,g,i,k), and the proposed synergy-based DRL in the purple frames of (b,d,f,h,j,l).

References

    1. Billard A., Kragic D. Trends and Challenges in Robot Manipulation. Science. 2019;364:eaat8414. doi: 10.1126/science.aat8414. - DOI - PubMed
    1. Sturm J., Stachniss C., Burgard W. A Probabilistic Framework for Learning Kinematic Models of Articulated Objects. JAIR. 2011;41:477–526. doi: 10.1613/jair.3229. - DOI
    1. Kochan A. Shadow Delivers First Hand. Ind. Robot. 2005;32:15–16. doi: 10.1108/01439910510573237. - DOI
    1. Kumar V., Xu Z., Todorov E. Fast, Strong and Compliant Pneumatic Actuation for Dexterous Tendon-Driven Hands; Proceedings of the 2013 IEEE International Conference on Robotics and Automation; Karlsruhe, Germany. 6–10 May 2013; pp. 1512–1519.
    1. Deimel R., Brock O. A Novel Type of Compliant and Underactuated Robotic Hand for Dexterous Grasping. Int. J. Robot. Res. 2016;35:161–185. doi: 10.1177/0278364915592961. - DOI

LinkOut - more resources