Skill Learning using Temporal and Spatial Entropies
for Accurate Skill Acquisition
Sang Hyoung Lee, Gyung Nam Han, Il Hong Suh
†
, and Bum-Jae You
Abstract— In manipulation tasks, skills are usually modeled
using the continuous motion trajectories acquired in the task
space. The motion trajectories obtained from a human’s multi-
ple demonstrations can be broadly divided into four portions,
according to the spatial variations between the demonstrations
and the time spent in the demonstrations: the portions in
which a long/short time is spent, and those in which the spatial
variations are large/small. In these four portions, the portions
in which a long time is spent and the spatial variation is
small (e.g., passing a thread through the eye of a needle) are
usually modeled using a small number of parameters, even
if such portions represent the movement that is essential for
achieving the task. The reason for this is that these portions
are slightly changed in the task space as compared with
the other portions. In fact, such portions should be densely
modeled using more parameters (i.e., overfitting) to improve
the performance of the skill because the movements of those
portions must be accurately executed to achieve the task. In
this paper, we propose a method for adaptively fitting these
skills based on the temporal and the spatial entropies calculated
by a Gaussian mixture model. We found that it is possible to
retrieve accurate motion trajectories as compared with those
of well-fitted models, whereas the estimation performance is
generally higher than that of an overfitted model. To validate
our proposed method, we present the experimental results and
evaluations when using a robot arm that performed two tasks.
I. I NTRODUCTION
In manipulation tasks, skills are usually learned using
continuous motion trajectories. An intelligent robot should
therefore be able to learn such skills, using a set of the
continuous motion trajectories obtained from a human’s mul-
tiple demonstrations. In such a set of motion trajectories, the
trajectories can be categorized into four portions, according
to the spatial variations between the demonstrations and
the duration (i.e., time spent) of the demonstrations. Let us
consider an example to intuitively understand the portions
that constitute a manipulation task. A robot learns a skill
for painting an assembly part based on a human’s multiple
demonstrations. The procedure is as follows: the robot first
lifts up a brush to the part. Next, it uses the brush to paint
*This work was supported by the Global Frontier R&D Program on
<Human-centered Interaction for Coexistence> funded by the National Re-
search Foundation of Korea grant funded by the Korean Government(MEST)
(NRF-M1AXA003-2011-0028353)
S. H. LEE is with the Department of Electronics and Computer Engi-
neering, Hanyang University, Seoul, Korea. zelog@hanyang.ac.kr
G. N. Han is with the Department of Electronics and Computer Engineer-
ing, Hanyang University, Seoul, Korea. mudian@hanyang.ac.kr
†
I. H. Suh is with the Department of Computer Science and Engineer-
ing, Hanyang University, Seoul, Korea. ihsuh@hanyang.ac.kr, All
correspondence should be addressed to I. H. Suh.
B. J. You is with Korea Institute of Science and Technology, Seoul, Korea.
ybj@kist.re.kr
the part along a zigzag path. Finally, the brush is withdrawn
from the part. In this painting example, Fig. 1 shows the
set of the motion trajectories obtained from a human’s
multiple demonstrations. In the figure, portion (1) involves
the painting of the assembly part, and the other portions
involve the lifting of the brush and withdrawal of the brush
from the part. Portion (1) takes up the longest time, although
the movement is slightly changed in the demonstrations when
the part is fixed in a specific location, as shown in Fig. 1.
Next, let us consider modeling a Gaussian Mixture Model
(GMM) using the motion trajectories in Fig. 1. The GMM
is estimated using Bayesian Information Criterion (BIC)
and Expectation-Maximization (EM) algorithms, as shown
in Fig. 2-(a). In the GMM, portion (1) in Fig. 1 is sparsely
modeled, even though the GMM is well fitted by the BIC
and EM algorithms. The problem can be formulated in terms
of the differences between Fig. 1 and Fig. 2-(b). There is
no zigzag path in the motion trajectories retrieved by a
Gaussian Mixture Regression (GMR) process. The GMM
should be modeled to use more parameters to retrieve the
zigzag path because this path is essential in achieving the
task. When the number of parameters (i.e., the number of
Gaussians) is forcefully increased, in this context, the rest of
the portions are more densely overfitted than portion (1), as
shown in Fig. 3. The reason for this is that the changes of
the movements in the rest of the portions are larger than they
are in portion (1). Although in portion (1), slightly accurate
motion trajectories can be retrieved by the GMM in Fig. 3,
the estimation performance is lower than that of the GMM
in Fig. 2. As shown in Fig. 3-(b), the motion trajectories
are overfitted in all portions. The model should be able to
remodeled using more parameters only in portion (1), while
maintaining the rest of the portions.
In this context, we propose a method for learning skills
that consider the spatial variations between multiple demon-
strations and the time spent in these demonstrations, based
on the entropies involved. Fig. 4 shows the entire process
of the proposed method. The set of the continuous mo-
tion trajectories obtained from multiple demonstrations is
temporally aligned using a Dynamic Time Warping (DTW)
algorithm, as shown in Fig. 4-(a). The motion trajectories
are projected onto the reduced dimensional subspace by
Principal Component Analysis (PCA), as shown in Fig. 4-
(b). A GMM is estimated to contain the temporal and the
spatial information based on the BIC and EM algorithms for
the calculation of the entropies of each portion, as shown
in Fig. 4-(c). The temporal and the spatial entropies are
calculated as per the Gaussians from the estimated GMM,
2013 IEEE International Conference on Robotics and Automation (ICRA)
Karlsruhe, Germany, May 6-10, 2013
978-1-4673-5642-8/13/$31.00 ©2013 IEEE 1315