Skip to content

learning and control from demonstration using LSTM and hybrid impedance controller

Notifications You must be signed in to change notification settings

nwilliterate/robot_writing_task

Repository files navigation

robot manipulator writing task

Authors: Seonghyeon Jo([email protected])

Date: Jan, 1, 2022

With the recent advances in robotics, robots are needed not only in the manufacturing field but also in various service fields. In order to realize a robot in the service fields, it is necessary to perform complex tasks and various work environments. Therefore traditional robots must be capable of easily performing physical interaction tasks. In this respect, learning from demonstration (LfD), a method in which robots acquire mimic skills by learning to imitate an expert, is an alternative way. LfD method learns the dynamics of the trajectory by optimizing it with a mathematical linear model. However, complex or overlapping patterns are difficult to learn and require the use of mixed models when learning position and force. In this paper we approach the problem of learning time-series data by predicting the output for the next state with the input for the current and previous states. Also, in terms of work, it is important to utilize the information of position/force so that the robot can perform its role well. In LfD, the imitation performance varies depending on the position/force controller. In this paper, we implicitly learn the policy for the task condition from the expert's demonstration and perform position/force-based hybrid impedance control that is not restricted from motion in the trajectory derived from the policy. In addition, it uses a position/force-based hybrid impedance controller to simultaneously learn position and force through kinematic teaching from experts to perform free movement. The proposed LfD method was experimentally verified by applying a writing task to a 7-DoF manipulator. The proposed method can be applied quickly and easily to position/force-based tasks without constraints.

video

https://www.youtube.com/embed/IUuEeZEsJNY?autoplay=1&mute=1&loop=1

Hnet com-image

Hnet com-image (1)

0. Raw data

1

1

1. preprocessing data

1

1

2. learning from demonstartion using LSTM network

1

1

3. trajectory interpolation

1

1

1

4. simulation result

1

1

5. experiment result

1

1

In learning from demonstation, the output is the next state of the robot taught by the expert.

\X_{k+1} = \f(\X_{k}, \X_{k-1},\cdots , \X_{k-l})

\X_{k} = [P_x, P_y, P_z, R_x, R_y, R_z, R_w, F_x]

\begin{align}
{} ^{i-1} T_{i}  =& {rot}_{x,\alpha_{i-1}} {trans}_{x,a_{i-1}}{rot}_{z,\thata_{i}} {trans}_{z,d_{i}} \\
=& \left[{\begin{array}{ccc|c}\cos \theta _{i} & -\sin \theta _{i} & 0 & a_{i-1}\\ \sin \theta _{i}\cos \alpha _{i-1} & \cos \theta _{i}\cos \alpha _{i-1} & -\sin \alpha _{i-1} & -d_{i}\sin \alpha _{i-1}\\\sin \theta _{i}\sin \alpha _{i-1} & \cos \theta _{i}\sin \alpha _{i-1} & \cos \alpha _{i-1} & d_{i}\cos \alpha _{i-1}\\\hline 0 & 0 & 0 & 1\end{array}}\right]
\end{align}

\begin{align}
{} ^{0} T_{i} &= {} ^{0}T_{1}\ {} ^{1}T_{2}\ {} ^{2}T_{3}\ \cdots\ {} ^{i-1}T_{i} \nonumber\\
&=\left[{\begin{array}{ccc|c}& & & \\& {} ^{0}R_{i} & & {} ^{0}P_{i}\\& & & \\\hline 0& 0& 0& 1\end{array}}\right]
\end{align}

About

learning and control from demonstration using LSTM and hybrid impedance controller

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages