Next Article in Journal
Biomimetics with Trade-Offs
Next Article in Special Issue
Design of a Multi-Mode Mechanical Finger Based on Linkage and Tendon Fusion Transmission
Previous Article in Journal
Computational Study of Stiffness-Tuning Strategies in Anguilliform Fish
Previous Article in Special Issue
Hand Grasp Pose Prediction Based on Motion Prior Field
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A High-Efficient Reinforcement Learning Approach for Dexterous Manipulation

1
College of Mechanical Engineering, Beijing University of Science and Technology, Beijing 100083, China
2
College of Mechanical Engineering, Zhejiang University of Technology, Hangzhou 310023, China
3
Guangdong Provincial Key Laboratory of Robotics and Intelligent System, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
*
Author to whom correspondence should be addressed.
Biomimetics 2023, 8(2), 264; https://doi.org/10.3390/biomimetics8020264
Submission received: 23 April 2023 / Revised: 5 June 2023 / Accepted: 6 June 2023 / Published: 16 June 2023
(This article belongs to the Special Issue Bionic Robot Hand: Dexterous Manipulation and Robust Grasping)

Abstract

:
Robotic hands have the potential to perform complex tasks in unstructured environments owing to their bionic design, inspired by the most agile biological hand. However, the modeling, planning and control of dexterous hands remain unresolved, open challenges, resulting in the simple movements and relatively clumsy motions of current robotic end effectors. This paper proposed a dynamic model based on generative adversarial architecture to learn the state mode of the dexterous hand, reducing the model’s prediction error in long spans. An adaptive trajectory planning kernel was also developed to generate High-Value Area Trajectory (HVAT) data according to the control task and dynamic model, with adaptive trajectory adjustment achieved by changing the Levenberg–Marquardt (LM) coefficient and the linear searching coefficient. Furthermore, an improved Soft Actor–Critic (SAC) algorithm is designed by combining maximum entropy value iteration and HVAT value iteration. An experimental platform and simulation program were built to verify the proposed method with two manipulating tasks. The experimental results indicate that the proposed dexterous hand reinforcement learning algorithm has better training efficiency and requires fewer training samples to achieve quite satisfactory learning and control performance.

1. Introduction

Dexterous hands are considered to be the most complex and diverse end effector for robots. Compared to other end effectors, dexterous hands have more flexibility in grasping objects [1], especially irregular objects. However, performing excellent grasping and manipulation with dexterous hands remains a grand challenge.
Excellent grasping performance relies highly on interaction with the environment, where reinforcement learning algorithms have been applied for decades to solved countless sequential challenging problems [2]. In particular, deep reinforcement learning (DRL), a combination of reinforcement learning and deep neural networks, has achieved fruitful results in the field of deep learning [3,4] for robotics, such as robotic dogs, unmanned vehicles and humanoid robots [5]. High-level, domain-specific knowledge for robotic planning could be learned automatically. Moreover, the computational framework for robotic tasks and automation have been developed to facilitate the study and development of manipulation skills for robotic hand.
Transferring difficulties in robotics to reinforcement learning [6] has enabled robots to address previously intractable problems [7], and DRL has provided a powerful means of representing complex policies in high-dimensional environments [8,9]. A considerable number of scholars have reported their achievements in applications on real-world robots. For example, Tsurumine proposed a deep reinforcement learning network that combines the nature of smooth policy to enhance stability and efficiency for robotic cloth manipulation [10]. A path integration reinforcement learning algorithm was proposed to solve the control problem of tendon-wired robotic actuators without modeling the dexterous hand or environment, inspired by the human brain’s muscle synergy to obtain planning and control [11]. However, reinforcement learning in real-world robots is very time-consuming [12]. For instance, a model-free reinforcement learning algorithm applied to a three-finger dexterous hand system took seven hours of training time to complete various control tasks in the real world [13]. This slow training efficiency remains a stumbling obstacle to real-time robotic control. To accelerate the training speed, researchers have proposed various techniques. For example, a hierarchical planning method was proposed to achieve the movement of turning small balls by learning the best strategies for different levels in a decoupled manner [14]. Q-learning was applied to high-level discrete motions, and an improved path integration strategy was used to learn Dynamic Movement Primitives (DMPs) for low-level control. Adding a small amount of artificial demonstration data was found to be effective. In order to improve the efficiency of data use, a DPG-R algorithm was proposed based on the DDPG algorithm and Q-Learning algorithm. The efficiency of learning data samples was improved by separating the update frequency of the network from the environment [15]. Kumar designed a non-grasping five-finger dexterous hand by learning the local linear time-varying model of the dexterous hand and building a local trajectory planning controller for the model [16]. These controllers were first initialized with manual demonstration data. Moreover, the control strategy could be learned only through tactile and dexterous hand proprioceptive feedback without relying on the visual feedback of the object. Nagabandi proposed planning with a deep dynamics model (PDDM), a model-based reinforcement learning algorithm [17]. In contrast to the local linear time-varying model, deep neural networks were used to learn a global model of the system, and the learned model was used to plan motion based on the cross-entropy method (CEM). The algorithm was applied to control the multi-fingered dexterous hand to perform multiple tasks [18,19]. Andrychowicz from the OpenAI company proposed a technique that uses domain randomization to pre-train dexterous hands in the virtual world through reinforcement learning. The pre-trained model was then fine-tuned using transfer learning to operate on a physical robotic hand. To estimate object poses from visual information, multi-layer convolutional neural networks were commonly used [20]. Proximal policy optimization (PPO) was often employed to train dexterous hands in a virtual environment using thousands of different parameters [21]. However, one difficulty with PPO is that it may require a large amount of training data to achieve acceptable results.
To address this issue, Soft Actor–Critic (SAC) was proposed based on the maximum entropy theory. Compared to algorithms such as Deep Deterministic Policy Gradient (DDPG), SAC is more efficient in selecting initial network parameters. Additionally, due to the adoption of maximum entropy, SAC can achieve better control performance for different random seeds. Therefore, applying the SAC algorithm to the dexterous hand system to improve data learning efficiency is a promising approach. Overall, in applying domain randomization, transfer learning, and multi-layer convolutional neural networks, SAC can improve the efficiency of dexterous hand learning in both virtual and real-world environments.
In this paper, a dynamic model of the dexterous hand is proposed based on generative adversarial architecture to learn the dynamic changing principle of the state of the dexterous hand. The contributions of this paper are as follows:
  • To address the problem of prediction error explosion over long spans, the state of the dexterous hand is learned through generative adversarial architecture with high prediction accuracy.
  • An adaptive trajectory planning method is proposed. Moreover, an adaptive trajectory programming kernel is built to generate High Value Area Trajectory (HVAT) based on the dexterous dynamic model and object. A smooth distance loss function and a U-shaped loss function are designed to calculate the loss value of the dexterous hand system at the reference point.
  • A new Actor–Critic-based reinforcement learning algorithm is proposed for the control of the dexterous hand.
This work enables the dexterous hand to explore high-value areas through a certain execution sequence, thereby improving the learning speed and performance of reinforcement learning algorithms.

2. Methodologies

The hybrid reinforcement learning strategy proposed in this paper is shown in Figure 1, which mainly consists of the SAC network, physical model of the dexterous hand and trajectory planning kernel. There are three networks in SAC, including CriticNet, ValueNet and ActorNet. Firstly, the system state obtained by the sensor in the manipulating environment is taken as the input of the SAC algorithm to the A c t o r ϕ network. Then the output motion of the dexterous hand is generated by trajectory planning. As the dexterous hand executes the motion, the environment state is updated and employed to renew A c t o r ϕ to generate the subsequent motion.

2.1. Generative Adversarial Architecture

In this study, the dexterous manipulation scenario is composed of a robotic hand and its operating target, which is a rotatable object such as a ball or blade. As shown in Figure 2, the state of the system includes variables such as the joint angle, joint velocity, joint torque, fingertip position, fingertip velocity, the position and velocity of the object, etc.
The state vector s = s 0 , , s N T and output vector o are established for the kinematics and dynamics of the dexterous hand:
s = f s , a
where a is the input of the system.
Traditionally, the dynamic model of the system is established through analyzing the dynamic model. In some cases of collision and slide, the dynamic model is hard to establish accurately. In this paper, a deep neural network is used to learn the dynamic model of the dexterous hand. The state equation of the dynamic model is able to automatically be designed. Another advantage is that the D y n N N is applied to obtain the changing principle of the system state. Therefore, when selecting the system state s , it only needs to satisfy that the state s satisfies the model of the Markov decision, instead of mapping the system state s into the observed value in traditional control methods.
The dynamic model of the system is obtained through the data-fitting method. Moreover, it can be used for model predictive control. Finally, a dynamic model of the system can be obtained:
s t + 1 = f s t , a t D y n N N s t , a t
where D y n N N represents a neural network model.
An improved generative adversarial nets (GAN) architecture is proposed for building dexterous hand dynamic models. The improved generative adversarial architecture consists of two modules, a generative network module and adversarial network module D i s c ϵ , which is represented by the network parameters ϑ and ϵ , respectively. The structure of D y n ϑ is shown in Figure 3.
The input of D y n ϑ is a two-dimensional vector s t a t , composed of the state of the dexterous hand at t-moment and the input of t-moment system. The output s t + 1 of D y n ϑ is the difference between the predicted state s t + 1 of the dexterous hand at t-moment and the system state s t at t-moment.
When predicting the state of the dexterous hand system at the next moment, the current tensor s t , a t T first passes through a one-dimensional convolution layer with a kernel size of 1, stride of 1 and out channels of 16. This convolutional layer is mainly used to extract the polarity characteristics of each element in s t , a t T , and the parameters of this convolutional layer are initialized to a normal distribution with a mean of 0 and variance of 1 / 8 2 . The tensor then passes through a linear and ReLu layer of 256 neurons for nonlinear mapping. Then the tensor continues to pass through a one-dimensional convolutional layer with a kernel size of 1, stride of 1 and out channels of 16. The parameters of this convolutional layer are initialized in the same way as those of the above convolutional layer, and are also used to extract the polar features of each element in the current layer. A basic microarchitecture of D y n ϑ is constructed by adjusting the polarity features of the extracted tensor and the number of nonlinear mapping levels. Theoretically, the more times the microarchitecture is reused, the more complex the dynamic model that can be learned by the D y n ϑ model. With the above structure, the final tensor passes through three mapping layers composed of linear and ReLu layers, and outputs the predicted value s t + 1 .
The adversarial network module D y n ϑ structure is shown in Figure 4. The D i s c ϵ donates the dexterous hand system state and the predicted dexterous hand system state by True or False, respectively. If the input D i s c ϵ of the tensor is a true dexterous hand state, then the D i s c ϵ should be 1, or True. On contrary, if the tensor entered by the D i s c ϵ is predicted by D y n ϑ , the output of the D i s c ϵ should be 0, or False. In practice, the D i s c ϵ outputs a value of 0 , 1 , and the larger the value, the more likely it is that the current input state is a real dexterous hand system state. D i s c ϵ also has the same structure of one-dimensional convolutional layers and nonlinear mapping as D y n ϑ for extracting polarity features.
In the dynamic model of a dexterous hand with a training model, the loss function of D y n ϑ is:
L D y n ( ϑ ) = E s t , a t , s t + 1 ~ B u f f e r [ χ log 1 D i s c ϵ D y n ϑ s t , a t + s t + 1 χ s t + 1 s t D y n ϑ s t , a t ]
The loss function of D i s c ϵ is:
L D i s c ϵ = E s t , a t ~ B u f f e r log 1 D i s c ϵ s t + log D i s c ϵ D y n ϑ s t , a t + s t
It can be concluded from the D y n ϑ network that the loss function consists of two parts, s t + 1 s t D y n ϑ s t , a t and log 1 D i s c ϵ D y n ϑ s t , a t + s t . The former is the prediction performance of the D y n ϑ , which reflects the difference of the real states before and after the system under motion control input a t . The latter can reflect the probability of the true state in the state of next moment s ^ t + 1 = D y n ϑ s t , a t + s t predicted by D i s c ϵ and D y n ϑ . For the D i s c ϵ network, the loss function also consists of two parts, respectively, the log 1 D i s c ϵ s t and log D i s c ϵ D y n ϑ s t , a t + s t . The former reflects the real state of the dexterous hand system for the possibility of True; and the latter reflects by the D y n ϑ prediction of the dexterous hand system status of the next moment s ^ t + 1 for the possibility of False. It is used for training D i s c ϵ by maximizing the loss function.

2.2. Theoretical Analysis of Trajectory Planning Algorithm

For a limited interval t = 0 : N , μ = a 0 , a 1 , , a N 1 represents the control input sequence. A loss function l s t , a t satisfying concave function is used to calculate the system loss value at t-moment. The loss function belongs to a part of the control system. The loss value is calculated according to the target trajectory and the corresponding system trajectory. Then the cumulative loss value of the system in the interval k = i : N 1 can be expressed as:
J i s , μ = i N 1 l s k , a k + l f s N
Afterwards, the optimal control problem can be expressed as finding an optimal trajectory sequence so that the loss accumulation value is a minimum within t = 0 : N :
μ s = a r g min μ J 0 s , μ
where μ represents the optimal motion sequence.
The loss value function V s , i = min μ i J i s , μ i , represents the cumulative value of system losses in the optimal motion subsequence μ i = a i , a i + 1 , , a N 1 , and it is worth noting that the size of V s , i only depends on the system state s . According to the memorylessness property, V s , i can be obtained:
V s , i = min a i l s i , a i + V s i + 1 , i + 1
It can also be obtained from s i + 1 = f ( s i , a i ) :
V s , i = min a i l s i , a i + V f s i , a i , i + 1
given an initial trajectory τ o l d : s 0 , a 0 , , s T at t = 0 : N . In order to obtain the optimal sequence of motion, the control contains two parts, dynamic programming and optimal control. The method of dynamic programming is backpropagation and forward propagation. The optimal control under the linear system is discussed below, namely f s t , a t = F s t a t . In nonlinear systems, optimal control is obtained by iteration and local linearization.
Applying backpropagation, the Jacobian matrix and Hessian matrix of value function can be obtained at t = i :
V s i = Q s Q a Q a a 1 Q a s
V s s i = Q s s Q s a Q a a 1 Q a s
Q t ,   K t ,   k t ,   V s t and V s s t in t = N : 0 can be calculated according to the above equations, and the expression of the motion increment δ a t with respect to the state change δ s t at each time interval of t = 0 : N is obtained. The motion sequence is applied to the forward propagation and the motion τ n e w is updated.
Applying forward propagation, the expressions of K i and k i at t = i are:
K i = Q a a 1 Q a s = l a a + f a T V s s + μ I f a + V s f a a 1 l a s + f a T V s s + μ I f s + V s f a s
k i = Q a a 1 Q a = l a a + f a T V s s + μ I f a + V s f a a 1 l a + f a T V s
The detail of the backpropagation and forward propagation are in the Supplementary Materials. Finally, in the nonlinear system, the optimal control is calculated by successive iterations, and each iteration includes the above backward propagation and forward propagation.

2.3. Design of Loss Function

The dexterous hand system is a nonlinear system. In order to enhance the stability of the controller by reducing the change gradient of the loss function, a smooth distance function is designed to evaluate the state loss of the dexterous hand system:
l s s = e 2 + α s 2 α s
where α s represents the smoothness parameter, as shown in Figure 5, and e = s s g o a l representants the system error. Choosing different loss function α s , the linearity will be regulated.
In this paper, α s = 0.2 is adopted, and l s and l s s are defined as follows:
l s = e e 2 + α s 2
l s s = e 2 + α s 2 1 2 e e 2 + α s 2 3 2
For the dexterous hand control input, it is necessary to ensure that the system control input has a small loss value within the allowable range. Moreover, after exceeding the system limit, the system loss value corresponding to the control input increases rapidly. In this paper, a U-shaped loss function is designed as follows:
l a a = α a 2 cosh a α a 1
where α a is a U-shaped coefficient.
When the control input is greater than 1 or less than −1, the loss value of the system feedback is too large. Figure 6 shows the image of the loss function for different α a . In this paper, α a = 0.3 is adopted. The loss function, l a and l a a are expressed as:
l a = α a sinh a α a
l a a = cosh a α a

2.4. Kernel Design of Adaptive Trajectory Planning

For nonlinear control systems, in order to obtain the optimal trajectory, there are three steps, as shown in Figure 7. Firstly, the reference point system model should be linearized, then the control feedback matrix K and k are obtained by backpropagation. As a result, the reference trajectory is obtained by forward propagation. An adaptive trajectory planning algorithm is built. By adjusting the linear search coefficient α and LM coefficient μ , the deviation can be controlled between the newly generated trajectory and the original reference trajectory. In the trajectory planning procedure, the existing control strategy is applied to generate an initial reference trajectory according to the initial system state. Then the program linearizes the reference points of the system at each moment of the reference trajectory with respect to D y n ϑ to obtain the local linear model, and obtains the Jacobian matrix and Hessian matrix of the loss function according to Equations (14), (15), (17) and (18).
Backpropagation is applied to calculate V s t ,   V s s t ,   Q t ,   K t ,   k t and V s N = l s ,   V s s N = l s s from time t = N: 0. A linear search is carried out. The linear search coefficient is initialized as α = 1 , and the new motion sequence A { a 0 , a 1 , , a N 1 } [22] is calculated by K t ,   k t to obtain the new trajectory. If the loss value obtained along the new system trajectory is less than that of the original trajectory, it means that the program accepts this iteration. However, if the loss value obtained along the trajectory is greater than that corresponding to the original system trajectory under the current coefficient, the change value is reduced by dividing the linear search coefficient α by 1.1. Moreover, the motion sequence is recalculated and the loss value is compared.
Then, if the linear search coefficient α has reached the minimum value, the LM coefficient μ is increased and K t , k t are recalculated to impose greater constraints on the deviation from the original trajectory. Afterwards, the linear search link is re-activated. The process is repeated until a new trajectory is adopted. By adjusting the linear search coefficient α and LM coefficient μ , the adaptive planning trajectory is generated.

2.5. Improved SAC Algorithm

In this paper, an improved SAC algorithm is proposed to generate the motion of the dexterous hand, including V a l u e ψ and T a r g e t V a l u e ψ s t networks to evaluate the value of motions in the current state and A c t o r ϕ network. The improved Bellman equation is obtained by incorporating the entropy of the control policy action distribution under the current system state s_t.
Q s t , a t = r s t , a t + γ E s t + 1 ~ p s t + 1 s t , a t V s t + 1
The data cache DataBuffer has an area for HVAT to store HVAT data generated through trajectory planning. In addition, this part of the data is generated by trajectory planning in the fitted dynamic model of the dexterous hand rather than obtained in the actual dexterous hand environment by the control strategy π . The loss functions of C r i t i c θ and V a l u e ψ under HVAT are obtained as follows:
J C r i t i c H V A T θ = E s t , a t , E U C T N t ~ τ _ B u f f e r c o n f t 2 ( C r i t i c θ s t , a t Q ^ s t , a t ) 2
where c o n f t is the confidence value predicted by D y n ϑ under s t , a t ,   c o n f = 1 E U C T N t D y n ϑ . E U C T N represents the cognitive uncertainty and can be written as:
E U C T N t D y n ϑ i E S s t + 1 i s t + 1 _ 2 E S
where E S = 3 is the model number of the ensemble, s t + 1 is the output of the D y n ϑ and s t + 1 _ is the average predicted value.
The loss function of V a l u e ψ can be written as:
J V a l u e H V A T ψ = E s t , a t ~ H V A T _ B u f f e r [ 1 2 V a l u e ψ s t 1 2 E a t ~ π C r i t i c θ s t , a t + C r i t i c θ s t , a t log π a t s t 2 ]
where a t is sampled in the HVAT area. So the evaluation scope of V a l u e ψ includes the motion distribution space generated by trajectory planning.
At this point, the loss function gradient for C r i t i c θ and V a l u e ψ is defined as:
^ θ J C r i t i c H V A T θ = c o n f t θ C r i t i c θ s t , a t C r i t i c θ s t , a t r s t , a t γ T a r g e t V a l u e ψ s t + 1
^ ψ J V a l u e H V A T ψ = ψ V a l u e ψ s t V a l u e ψ s t 1 2 E a t ~ π C r i t i c θ s t , a t + C r i t i c θ s t , a t log π a t s t
Different from the direct generation of the trajectory by the control strategy, this method makes the high-value state region updated preferentially in the value iteration; thus, it indirectly accelerates the learning of the control strategy and avoids the direct influence of the low-value trajectory on the updating strategy of control.

Iteration of Control Policies

To maximize the entropy of actor–critic, the updating rule for the control policy π is defined as:
π n e w = arg min ϕ D K L π s t e x p Q π o l d s t , Z π o l d s t
where Z π o l d ( s t ) is the sum of all e x p ( Q π o l d ( s t , ) ) .
Since Critic is used to denote Q , the loss function for π ϕ can be expressed as follows:
J π ϕ = E s t ~ B u f f e r D K L π ϕ s t e x p C r i t i c θ s t , Z θ s t
Meanwhile, A c t o r ϕ ϵ t ; s t neural network is applied for reparameterization:
a t = A c t o r ϕ ϵ t ; s t
where ϵ t is the noise vector sampled from the Gaussian distribution.
Therefore, Equation (24) can be rewritten as:
J π ϕ = E s t ~ B u f f e r , ϵ t ~ N log π ϕ A c t o r ϕ ϵ t ; s t s t C r i t i c θ s t , A c t o r ϕ ϵ t ; s t
It is worth noting that A c t o r ϕ identifies a unique control strategy π ϕ . Its gradient can be approximated as:
^ ϕ J π ϕ = ϕ log π ϕ a t s t + a t log π ϕ a t s t a t C r i t i c θ s t , a t ϕ A c t o r ϕ ϵ t ; s t
Finally, the parameters of the A c t o r ϕ network are updated using gradient descent:
ϕ ϕ λ π ^ ϕ J π ϕ
where λ π is the learning rate of the control strategy π .

3. Experiments

3.1. Model-Fitting Experiments for Dynamic Processes

The dynamic model-fitting experiment of the dexterous hand is performed in a customized two-fingered hand simulation environment, as shown in Figure 8.
The hyperparameters used in this experiment are shown in Table 1 and Table 2. The cycle in the experimental simulation environment is 0.02 s, and the time step used in each simulation is 500.
We designed two types of prediction errors as performance indexes for the dynamic model of the dexterous hand based on the number of predicted steps; that is, short-span and long-span prediction errors. These indexes are used to analyze the variation principle of prediction errors and the phenomenon of error explosion with an increase in learning times. To do so, we initialized the dexterous hand system status and used the same control strategy for both types of prediction errors. The short-span prediction error is calculated based on the dynamic model and is obtained by predicting the state change of the dexterous hand system according to the reference trajectory in the next moment. The calculation for the short-span prediction error is:
δ = m e a n E s t , a t ~ τ i = 0 T 1 s i + 1 s ^ i + 1
where s ^ i + 1 = s i D y n ϑ s i , a i .
The long-span prediction error is used to predict the complete state trajectory of the dexterous hand system based on its initial state and motion sequence, and the prediction error is calculated by comparing it with the reference state trajectory. Unlike the short-span prediction error, the long-span prediction error can better reflect the model’s ability to learn the overall trend of the system state. The calculation for the long-span prediction error is:
δ = m e a n E s 0 , s t , a t ~ τ i = 0 T 1 s i + 1 s ^ i + 1
s ^ i + 1 = s i + D y n ϑ ( s ^ i , a i ) s ^ i = s ^ i 1 + D y n ϑ s ^ i 1 , a i 1 s ^ 0 = s 0
Then, the fixed state of the transfer dataset was utilized to train the dynamic model of the dexterous hand. Initially, the SAC algorithm was employed to generate motion sequence, and the control strategy was updated according to the control task to obtain the system state transfer dataset, which includes 10,000 system state transfer tuples. The dataset was then used to train the two models the same number of times. After each training, the prediction performance was tested five times, and the average value was computed to obtain the model’s prediction error. Moreover, after each training, the dynamic model’s two different span prediction errors, with and without the generative adversarial architecture, were recorded, and the error trend chart was obtained, as shown in Figure 9. In the figure, the horizontal direction represents an increase in the predicted time, the vertical direction represents the state size of the dexterous hand system and the red curve represents the real system state, while the blue curve represents the system state predicted by the dynamic model.
Figure 9a illustrates the trend of short- and long-span prediction errors in the dynamic model with increasing training times. The red curve indicates that the generative adversarial architecture was adopted, while the blue curve indicates that the architecture was not adopted. The smaller the value in the figure, the higher the prediction accuracy. The lower part of the figure is divided into four subgraphs, respectively, based on the dynamic model of generative adversarial architecture in training, to reach 50(I), 100(II), 150(III), 200 times(IV) of full long-span prediction, forecasting the state change curve and the true state graph.
As can be seen in Figure 9, with the increase in the number of dynamic model trainings, the overall prediction accuracy of the model for the state change of the environment was gradually improved. The model without the generative adversarial architecture was gradually stabilized after 120 trainings on the short-span prediction error, which is slightly better than that using the generative adversarial architecture mode. The short-span prediction error using the generative adversarial architecture model shows an upward trend at the initial stage of training, and eventually falls back to a smaller value, which is due to the fact that the generative adversarial architecture was used for training. Because the loss function of D y n ϑ in Equation (3) contains adversarial loss items, the error value in the graph experienced an upward and then a downward trend during the training progress.
Regarding long-span prediction performance, the use of the generative adversarial architecture significantly reduced the prediction error compared to the model without the architecture. The prediction error of the former eventually stabilized at the same order of magnitude as the short-span prediction error, while the latter exhibits the phenomenon of error explosion. These results demonstrate that the dynamic model using the generative adversarial architecture can capture the overall trend of the system state change without significant loss in short-span prediction performance. Thus, when the control strategy remains unchanged, the prediction performance of the model for long spans can be significantly improved, thereby enhancing the model’s overall prediction accuracy.

3.2. Adaptive Trajectory Planning Experiments of the Dexterous Hand

In order to verify the performance of the trajectory planning kernel, experiments with the dexterous hand were conducted. The experimental hyperparameters are listed in Table 3:
In order to reduce the computational complexity, the nimble fingers space was selected as an important condition. A total of 100 rounds of dexterous hands’ state transfer data were collected firstly, with a total number of 100 steps for each round. Then, the dynamic model was fitted. Moreover, the dynamic model and the trajectory planning kernel were used for model predictive control. Then, 10 steps were predicted forward at each time, and the spatial coordinates of the fingertips of dexterous hands were randomly generated as target positions for trajectory planning experiments.
Figure 10 shows the fingertip trajectory of the dexterous hand under a total of 400 times of dynamic model training and under different training times with the dynamic model, where green and red represent finger 1 and finger 2, respectively, the dot represents the initial coordinate of the fingertip, the asterisk is the target coordinate of the fingertip and the curve is the actual trajectory of the corresponding fingertip. Figure 9 selected the dynamic model under different training times using adaptive kernel trajectory planning; in the fingertip trajectory graph, we can see that with the increase in dynamic model training, at the same time and steps, the tip of the finger position with respect to the target movement shows a trend of gradually ascending, and the trajectory planning effect also was improved.
According to the experiment, in the early stage of the dynamic model training, due to the large cognitive deviation of the model in the state change of the environment, the dynamic model error occurred in the fingertip trajectory planning near the start point, which led the trajectory planning into local minimum, wrong direction and other problems, resulting in the fingertip’s mismatch with the desired position. Although the method of model predictive control can reduce the trajectory error caused by the dynamic model to a certain extent, the performance of the dynamic model had a major impact on the trajectory planning within a certain number of control steps. In the later stages of dynamic model training, a high-value fingertip trajectory, namely HVAT data, could be obtained through adaptive trajectory planning due to the improved accuracy of the model.

3.3. Controlling Experiments of the Dexterous Hand

The HVAT data are obtained via the above method and the dexterous hand control experiment was carried out by combining the SAC algorithm. The pseudo-code of the specific algorithm in the experiment is shown in Table 4.
The experimental platform is shown in Figure 11, which includes a two-fingered hand and its sensing and controlling units, a camera and the ROS program.
The experimental results of fingertip controlling are shown in Figure 12, in which the algorithm proposed in this paper (the green one) is compared with the model-free reinforcement learning algorithms SAC [22], PPO [23], DDPG [24], Policy Gradient [24,25,26] and Actor–Critic [25]. A total of 250 sets of training, which contain 5000 simulation time units, are adopted. From Figure 12, it can be concluded that the exploration ability of the model-free reinforcement learning algorithm is improved. By applying the maximum entropy control strategy, the experimental curve can achieve the maximum reward in the shortest time. Therefore, the overall learning speed of the algorithm is faster than that of the others.
For the task of rotating the pointer, the proposed control algorithm achieved the expected motion to a specified angle locally on the experimental platform by learning the relationship between the dexterous hand movements and the state changes of the pointer. Figure 13 shows the captured image by the camera of the experimental platform during the experiment, in which the blue numbers at the top of the image represent the randomly generated desired angle and the white numbers represent the current angle of the pointer. It can be seen from the photos that the dexterous hand can change the angle of the pointer by touching it. Figure 14 shows the curve of the angle changes of the pointer in this experiment.
Meanwhile, comparative experiments were conducted on different reinforcement learning algorithms for the same experimental environment, as shown in Table 5, which indicates the number of sample collection times required by different algorithms to achieve the same reward value. Compared with the others, it can be found that the control algorithm proposed in this paper requires fewer data samples to achieve the same control effect; i.e., the average reward value is in the same range in each round of the control experiment. It should be noted that the reward size of each round is related to the duration of the control steps. The control time steps of each round are set to be 5000.
The experimental results once again proved that the model-based reinforcement learning algorithm proposed in this paper for dexterous hand control can significantly improve the efficiency of training and enhance the learning speed compared to other model-free reinforcement learning algorithms.

4. Conclusions

Controlling both dexterous hand motion and in-hand manipulation in an unstructured environment with multiple tasks remains a great challenge. To address this issue, one feasible solution is to allow dexterous hands to interact with the environment and learn to achieve control goals through interaction, similar to that of humans.
In this paper, we proposed a dynamic model of dexterous hands based on generative adversarial architecture, which is capable of learning the dynamic change principle of the dexterous hand system state. To solve the problem of prediction error explosion over long spans, the generative adversarial architecture was employed to learn the entire change rule of the dexterous hand system state while maintaining local prediction accuracy. Then, an iterative optimal control theory of trajectory planning was proposed. An adaptive trajectory planning kernel was built based on this theory to generate the HVAT trajectory according to the dynamic model of the dexterous hand. The kernel can adjust the linear search coefficient α and LM coefficient μ to constrain the change of trajectory, so as to achieve the purpose of self-adaptation. Subsequently, a U-shaped loss function was designed to calculate the loss value of the dexterous hand system at the reference point. Finally, the SAC algorithm was improved and a new Actor–Critic-based reinforcement learning algorithm was proposed for the end-to-end control of dexterous hands, including two critic networks for evaluating the value of motions in the current state. A V a l u e ψ and T a r g e t V a l u e ψ s t network was used to assess the value of the current state, and an A c t o r ϕ network was used to generate motions. For random data in the cache, the algorithm adopts the update strategy of the maximum entropy framework, while for HVAT data in the cache, the algorithm adopted the hybrid update strategy.
The proposed algorithm can be used to enable dexterous hands to explore in high-value areas through a certain execution sequence, thereby improving the learning speed and performance of reinforcement learning algorithms. The dexterous hand was enabled to explore high-value areas through a specific execution sequence, thereby improving learning speed and performance. The above contributions provide a robust framework for end-to-end control of dexterous hands in unstructured environments with multitasking capabilities.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/biomimetics8020264/s1. S1: Backpropagation; S2: Forward propagation.

Author Contributions

Conceptualization, J.Z. (Jianhua Zhang) and G.B.; methodology, J.Z. (Jinyu Zhou); software, X.Z.; validation, J.Z. (Jinyu Zhou) and X.Z.; formal analysis, G.L.; data curation, S.C.; writing—original draft preparation, X.Z. and S.Q.; writing—review and editing, G.L. All authors have read and agreed to the published version of the manuscript.

Funding

The research was financially supported by the Joint Fund of the National Natural Science Foundation of China with Shenzhen City (Grant No. U2013212), the Key Research and Development Program of Zhejiang (Grant No. 2021C04015) and the Natural Science Foundation of Zhejiang (Grant No. LZ23E050005).

Data Availability Statement

There are no data to be shared.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Gao, G.; Gorjup, G.; Yu, R.B.; Jarvis, P.; Liarokapis, M. Modular, Accessible, Sensorized Objects for Evaluating the Grasping and Manipulation Capabilities of Grippers and Hands. IEEE Robot. Autom. Lett. 2020, 5, 6105–6112. [Google Scholar] [CrossRef]
  2. Nguyen, T.T.; Nguyen, N.D.; Nahavandi, S. Deep Reinforcement Learning for Multiagent Systems: A Review of Challenges, Solutions, and Applications. IEEE Trans. Cybern. 2020, 50, 3826–3839. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Arulkumaran, K.; Deisenroth, M.P.; Brundage, M.; Bharath, A.A. Deep reinforcement learning: A brief survey. IEEE Signal Process. Mag. 2017, 34, 26–38. [Google Scholar] [CrossRef] [Green Version]
  4. Li, X.; Serlin, Z.; Yang, G.; Belta, C. A formal methods approach to interpretable reinforcement learning for robotic planning. Sci. Robot. 2019, 4, eaay6276. [Google Scholar] [CrossRef] [PubMed]
  5. Kormushev, P.; Calinon, S.; Caldwell, D.G. Reinforcement learning in robotics: Applications and real-world challenges. Robotics 2013, 2, 122–148. [Google Scholar] [CrossRef] [Green Version]
  6. Iwata, T.; Shibuya, T. Adaptive modular reinforcement learning for robot controlled in multiple environments. IEEE Access 2021, 9, 103032–103043. [Google Scholar] [CrossRef]
  7. Wiedemann, T.; Vlaicu, C.; Josifovski, J.; Viseras, A. Robotic Information Gathering With Reinforcement Learning Assisted by Domain Knowledge: An Application to Gas Source Localization. IEEE Access 2021, 9, 13159–13172. [Google Scholar] [CrossRef]
  8. Kober, J.; Bagnell, J.A.; Peters, J. Reinforcement learning in robotics: A survey. Int. J. Robot. Res. 2013, 32, 1238–1274. [Google Scholar] [CrossRef] [Green Version]
  9. Kiran, B.R.; Sobh, I.; Talpaert, V.; Mannion, P.; Al Sallab, A.A.; Yogamani, S.; Perez, P. Deep Reinforcement Learning for Autonomous Driving: A Survey. IEEE Trans. Intell. Transp. Syst. 2022, 23, 4909–4926. [Google Scholar] [CrossRef]
  10. Tsurumine, Y.; Cui, Y.; Uchibe, E.; Matsubara, T. Deep reinforcement learning with smooth policy update: Application to robotic cloth manipulation. Robot. Auton. Syst. 2019, 112, 72–83. [Google Scholar] [CrossRef]
  11. Rombokas, E.; Malhotra, M.; Theodorou, E.A.; Todorov, E.; Matsuoka, Y. Reinforcement Learning and Synergistic Control of the ACT Hand. IEEE-Asme Trans. Mechatron. 2013, 18, 569–577. [Google Scholar] [CrossRef] [Green Version]
  12. Rothmann, M.; Porrmann, M. A Survey of Domain-Specific Architectures for Reinforcement Learning. IEEE Access 2022, 10, 13753–13767. [Google Scholar] [CrossRef]
  13. Zhu, H.; Gupta, A.; Rajeswaran, A.; Levine, S.; Kumar, V. Dexterous Manipulation with Deep Reinforcement Learning: Efficient, General, and Low-Cost. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019. [Google Scholar]
  14. Andrychowicz, M.; Baker, B.; Chociej, M.; Jozefowicz, R.; McGrew, B.; Pachocki, J.; Petron, A.; Plappert, M.; Powell, G.; Ray, A.; et al. Learning dexterous in-hand manipulation. Int. J. Robot. Res. 2020, 39, 3–20. [Google Scholar] [CrossRef] [Green Version]
  15. Gao, Y.; Ma, S.W.; Liu, J.J.; Xiu, X.C. Fusion-UDCGAN: Multifocus Image Fusion via a U-Type Densely Connected Generation Adversarial Network. IEEE Trans. Instrum. Meas. 2022, 71, 1–13. [Google Scholar] [CrossRef]
  16. Jain, D.; Li, A.; Singhal, S.; Rajeswaran, A.; Kumar, V.; Todorov, E. Learning Deep Visuomotor Policies for Dexterous Hand Manipulation. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Montreal, CA, Canada, 20–24 May 2019; pp. 3636–3643. [Google Scholar]
  17. Fu, Q.; Santello, M. Chapter 3—Sensorimotor Learning of Dexterous Manipulation. In Human Inspired Dexterity in Robotic Manipulation; Watanabe, T., Harada, K., Tada, M., Eds.; Academic Press: Cambridge, MA, USA, 2018; pp. 27–52. [Google Scholar]
  18. Kim, C.I.; Kim, M.; Jung, S.; Hwang, E. Simplified Frechet Distance for Generative Adversarial Nets. Sensors 2020, 20, 1548. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Luo, W.J.; Wang, P.; Wang, J.H.; An, W. The research process of generative adversarial networks. In Proceedings of the International Seminar on Computer Science and Engineering Technology (SCSET), Shanghai, China, 17–18 December 2018. [Google Scholar]
  20. Abdelgader, A.A.; Viriri, S. Deep Residual Learning for Human Identification Based on Facial Landmarks. In Proceedings of the 15th International Work-Conference on Artificial Neural Networks (IWANN), Gran Canaria, Spain, 12–14 June 2019; pp. 61–72. [Google Scholar]
  21. Xu, Z.; Zhou, X.; Bai, X.; Li, C.; Chen, J.; Ni, Y. Attacking asymmetric cryptosystem based on phase truncated Fourier fransform by deep learning. Acta Phys. Sin. 2021, 70, 144202. [Google Scholar] [CrossRef]
  22. Dolanc, G.; Strmčnik, S. Design of a nonlinear controller based on a piecewise-linear Hammerstein model. Syst. Control Lett. 2008, 57, 332–339. [Google Scholar] [CrossRef]
  23. Hamalainen, P.; Babadi, A.; Xiaoxiao, M.; Lehtinen, J. PPO-CMA: Proximal Policy Optimization with Covariance Matrix Adaptation. In Proceedings of the 2020 IEEE 30th International Workshop on Machine Learning for Signal Processing (MLSP), Espoo, Finland, 21–24 September 2020; p. 6. [Google Scholar] [CrossRef]
  24. Bernat, J.; Apanasiewicz, D. Model Free DEAP Controller Learned by Reinforcement Learning DDPG Algorithm. In Proceedings of the 2020 IEEE Congreso Bienal de Argentina (ARGENCON), Resistencia, Argentina, 1–4 December 2020; p. 6. [Google Scholar] [CrossRef]
  25. de Jesus, J.C.; Kich, V.A.; Kolling, A.H.; Grando, R.B.; Cuadros, M.A.d.S.L.; Gamarra, D.F.T. Soft Actor-Critic for Navigation of Mobile Robots. J. Intell. Robot. Syst. 2021, 102, 31. [Google Scholar] [CrossRef]
  26. Ju-Seung, B.; Byungmoon, K.; Huamin, W. Proximal Policy Gradient: PPO with Policy Gradient. arXiv 2020, arXiv:2010.09933. [Google Scholar]
Figure 1. Schematic diagram of the proposed method.
Figure 1. Schematic diagram of the proposed method.
Biomimetics 08 00264 g001
Figure 2. Dynamic model of the dexterous hand.
Figure 2. Dynamic model of the dexterous hand.
Biomimetics 08 00264 g002
Figure 3. Generative network diagram for the dexterous hand.
Figure 3. Generative network diagram for the dexterous hand.
Biomimetics 08 00264 g003
Figure 4. Adversarial network diagram of the dexterous hand.
Figure 4. Adversarial network diagram of the dexterous hand.
Biomimetics 08 00264 g004
Figure 5. Loss function for different α s .
Figure 5. Loss function for different α s .
Biomimetics 08 00264 g005
Figure 6. U-shaped loss function for different α a .
Figure 6. U-shaped loss function for different α a .
Biomimetics 08 00264 g006
Figure 7. Steps for adaptive trajectory planning.
Figure 7. Steps for adaptive trajectory planning.
Biomimetics 08 00264 g007
Figure 8. Simulation environment of a two-fingered hand.
Figure 8. Simulation environment of a two-fingered hand.
Biomimetics 08 00264 g008
Figure 9. Evolution of the dynamic model’s prediction error.
Figure 9. Evolution of the dynamic model’s prediction error.
Biomimetics 08 00264 g009
Figure 10. The trajectory of the dexterous hand.
Figure 10. The trajectory of the dexterous hand.
Biomimetics 08 00264 g010aBiomimetics 08 00264 g010b
Figure 11. Scheme of the experimental platform.
Figure 11. Scheme of the experimental platform.
Biomimetics 08 00264 g011
Figure 12. Fingertip controlling experimental results.
Figure 12. Fingertip controlling experimental results.
Biomimetics 08 00264 g012
Figure 13. Rotation pointer experiments.
Figure 13. Rotation pointer experiments.
Biomimetics 08 00264 g013
Figure 14. Convergence process of the pointer orientation.
Figure 14. Convergence process of the pointer orientation.
Biomimetics 08 00264 g014
Table 1. Parameters of the improved SAC algorithm.
Table 1. Parameters of the improved SAC algorithm.
ParameterSymbolValue
Learning rate l r 0.0003
Hidden layer d e e p 2
Number of monolayer neurons 256
The entropy coefficient t m p 10
Discount factor γ 0.99
Batch size 256
Table 2. Parameters of the dynamic model of the dexterous hand.
Table 2. Parameters of the dynamic model of the dexterous hand.
ParameterSymbolValue
Learning rate l r 0.0001
Number of monolayer neurons 256
Number of polar layers 2
Adversarial coefficient χ 0.4
Weight decay 0.0001
Table 3. Kernel hyperparameters of adaptive trajectory planning.
Table 3. Kernel hyperparameters of adaptive trajectory planning.
ParameterSymbolValue
Initial LM coefficient μ 0 0.1
Maximum LM coefficient M A X μ 10
Gain coefficient 1.2
Initial gain coefficient 0 1.5
Maximum number of iterations 20
Predicted steps to control 15
Table 4. Reinforcement learning algorithm for dexterous hand manipulation.
Table 4. Reinforcement learning algorithm for dexterous hand manipulation.
01: Initialize these network parameters and DataBuffer, such as D y n ϑ , D i s c ϵ , A c t o r ϕ , C r i t i c θ , V a l u e ψ
02: Collect transfer data for the initial system state D a t a B u f f e r
03: Each episode executes in the iteration loop
04: Sample control targets randomly
05: The number of steps at each time t is executed in the iteration loop
06: Obtain s t , a t , s t + 1 , r t D a t a B u f f e r , with the A c t o r ϕ sampling motion
07: Policy updates in the iteration loop
08: Sample batch data from DataBuffer
09: Update the network parameters of C r i t i c θ , V a l u e ψ , A c t o r ϕ , T a r g e t V a l u e ψ s t
10: Dynamic model updates in the iteration loop:
11: Sample batch data from DataBuffer
12: Update D y n ϑ and D i s c ϵ
13: The HVAT sample is taken from the iteration loop
14: Sample the initial data from DataBuffer
15: The optimal trajectory planning is carried out to generate HVAT data D a t a B u f f e r
Table 5. Comparison of training sample efficiency of different algorithms.
Table 5. Comparison of training sample efficiency of different algorithms.
MethodologyReward-10,000Reward-8000Reward-6000Reward-4000
Ours6102025
SAC7122130
PPO135151162N.A.
DDPG71863N.A.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, J.; Zhou, X.; Zhou, J.; Qiu, S.; Liang, G.; Cai, S.; Bao, G. A High-Efficient Reinforcement Learning Approach for Dexterous Manipulation. Biomimetics 2023, 8, 264. https://doi.org/10.3390/biomimetics8020264

AMA Style

Zhang J, Zhou X, Zhou J, Qiu S, Liang G, Cai S, Bao G. A High-Efficient Reinforcement Learning Approach for Dexterous Manipulation. Biomimetics. 2023; 8(2):264. https://doi.org/10.3390/biomimetics8020264

Chicago/Turabian Style

Zhang, Jianhua, Xuanyi Zhou, Jinyu Zhou, Shiming Qiu, Guoyuan Liang, Shibo Cai, and Guanjun Bao. 2023. "A High-Efficient Reinforcement Learning Approach for Dexterous Manipulation" Biomimetics 8, no. 2: 264. https://doi.org/10.3390/biomimetics8020264

Article Metrics

Back to TopTop