Elsevier

Robotics and Autonomous Systems

Volume 87, January 2017, Pages 28-37
Robotics and Autonomous Systems

Gaussian Mixture Model for 3-DoF orientations

https://doi.org/10.1016/j.robot.2016.10.002Get rights and content

Abstract

This paper presents learning and generalization algorithms for Gaussian Mixture Model (GMM) in order to accurately encode 3-DoF orientations and Euclidean variables in a common model. We employ correct displacement, integration and weighted averaging arithmetics for unit quaternions to adapt the learning and generalization methods of standard GMMs.

We validate the proposed method in three different applications, learning a 3-dimensional rotation matrix, learning reachable space of a robot, and learning the motion model from demonstrations. We show good experimental results compared to the state-of-the-art method.

Introduction

We propose a modified learning algorithm for Gaussian Mixture Models (GMM) that treats the fundamentally non-Euclidean geometry of the space of 3D-orientations in a principled way, in order to encode both, 3D-orientations and standard Euclidean variables in a common model. GMM employs a mixture of Gaussian probability density functions to model the statistic distribution of a multi-variate dataset, and it has been successfully applied to many robotic applications. One group of applications is concerned with learning robot motions from human demonstration. In this context GMMs were successfully applied to model reaching and moving [1], [2], [3], hitting [4], [5] or catching [6], [7] tasks. Another group of applications handles estimating a distribution of admissible grasping postures for an object and modeling the reachable space of a robot, i.e. the possible positions and orientations the robot’s hand can achieve [7]. By combining both acquired models, a robot can calculate if any graspable part of the object lies inside its reachable space.

A fundamental issue arising when we try to apply standard GMMs to orientation representations, is the non-Euclidean nature of SO (3), the space of rotations in R3. Instead of being flat, the orientation space is bounded and curved: After a rotation of 360°, we are back at the origin. However, the formulation of the Gaussian probability distribution in standard GMMs assumes a Euclidean structure and correspondingly employs a Euclidean distance measure.

Most existing workexploits the fact that locally SO (3) is similar to R3. That is, for small rotation magnitudes, one can successfully employ standard Euclidean distances. For example, three of the four quaternion components [4] or scaled axis–angle representations [6] were used, without any modifications to the GMM learning algorithm.

In the same line of research, Feiten et al. [8], [9] introduced Mixtures of Projected Gaussians (MPG). MPG represents a distribution over SE (3), by a tangent point (as unit quaternions) and a Gaussian distribution over the projected space of orientation and translation. MPG provides outstanding performances in many researches [9], [10]. However, the projection of MPG [9] is unstable near the opposite orientation to the base orientation as shown in Fig. 1. Additionally, MPG doesn’t provide a regression method in the literature. Although MPG provides a way to fuse two projected Gaussians, the fusion is only valid when the angle between the normals of the tangent spaces is less than 15° [9]. Moreover, similar to the above researches [4], [6], they also exploit the fact that the projected space of SO (3) is locally similar to R3.

As these methods [4], [6], [9] use the Euclidean space EM [1], [11], [12] in order to model the non-Euclidean space dataset, they cannot exactly capture the distributions over SO (3). Hence it brings about a serious modeling error when it models a wide range of orientation distributions with a small number of Gaussians.

Beside the GMMs introduced above, there are other GMM extensions (such as Dirichlet process GMMs [13], Quantum GMMs [14], [15], etc.). However these methods also consider Euclidean space variables only, and are not suitable to model SO(3).

For large rotation magnitudes, Gaussian process (GP) with a dual quaternion was introduced by [16]. They models the distribution of 6D rigid body poses using GP, and the poses are represented by a dual quaternion. As a kernel function of GP, they employ the quaternion distance measure [17] in order to compute one-dimensional distances between unit quaternions. However, the regression process is computationally expensive for a large dataset, as it use all the training points, and it recomputes the covariance matrix of the training points and a query point (c.f. one of our example uses 107 training data points, see Section 3.2).

Another line of research for large rotation magnitudes, uses a more redundant, R6 representation, employing two columns of the corresponding 3 × 3 rotation matrix [7] (named TRGMM). Although it can model wide range orientations using standard GMMs, it has several limitations: more dimensions require more training samples for GMM learning; no measures are taken to ensure orthonormality of reconstructed orientation matrices; and the standard Euclidean distance between two orientations in R6 representation doesn’t exactly measure the distance in orientation space as shown in Fig. 1.

As a result of their fundamentally different structure, there exists no singularity-free, differentiable one-to-one mapping (a homeomorphism) between SO (3) and R3. Only if we consider higher-dimensional spaces, e.g. 3 × 3 rotation matrices, a smooth mapping exists. Additionally to a proper choice of representation, also some standard operations, e.g. the distance measure, need to be adapted in order to accurately handle the non-Euclidean structure of SO (3) in Gaussian distributions.

Alternatively, some direct methods were proposed to treatdistributions on the orientation manifold correctly, e.g. von Mises–Fisher [18], Bingham [19] and wrapped Gaussian [20] distributions.1 These [18], [19] are antipodally symmetric probabilitydistributions on an unit hypersphere. They can model the distribution of SO (3) orientations (as unit quaternions) on a 4D hypersphere. However, it is computationally expensive (especially the integral in the algorithm). Furthermore, composing the distribution of orientation with position was not considered. In this paper, we propose an alternative approach by adapting the learning procedure of GMM in order to intrinsically and accurately treat the geometry of a chosen orientation representation and to encode the distributions of orientation and Euclidean space variables in a common model. We have chosen unit quaternions to represent orientations, because they provide convenient mathematical notation and constitute a minimal, though four-dimensional, smooth representation of SO (3) [22].

In aEuclidean space, the distance between two data points is the norm of the distance vector—computed by subtracting one data point from the other. Averaging of data points (to obtain the mean) is accomplished by summing up all vectors and dividing by their number. However, in a non-Euclidean space like SO (3), these operations are not well-defined anymore. For example, the mean of two unit quaternions, q and q, yields the null vector, which is not a unit quaternion anymore.

In a few applications [17], [23], the quaternion distance issue is resolved by employing the quaternion logarithm. Ude et al. [17], [23] employ it in order to express the distance vector from an initial orientation to a goal orientation for learning control of complex robot behaviors with DMPs. To perform weighted quaternion averaging, Markley et al. [24] proposed a robust and accurate method based on PCA.

These motivate the two core contributions of the present paper: (1) An extension of the standard GMM equations that makes them consistent with the underlying non-Euclidean geometry of the space of 3D-orientations (we name it as unit quaternions GMM; in short QGMM). (2) A comparison of the new algorithm (QGMM) with standard GMM on three different problems in robotics, showing that the proper treatment of the geometry of 3D-orientation space can lead to significant accuracy gains.

The remainder of this paper is organized as follows. Section 2 presents the summary of quaternion arithmetics and modified learning procedures of GMM. In Section 3, we evaluate the method in three different learning applications. Finally, we conclude with a summary and remaining issues in Section 4.

Section snippets

GMM for unit quaternions

GMM learning makes use of some basic arithmetic operations, including displacement, displacement integration, and weighted averaging. To accurately represent unit quaternions in GMMs we need to replace the standard Euclidean realizations with appropriate replacement functions that apply to unit quaternions as summarized in Table 1. After introducing the definitions of these replacement functions in the next subsection, we will discuss the required modifications to the GMM algorithm.

Throughout

Evaluation

In order to illustrate the general nature of the proposed method and to evaluate some of its benefits as compared to a previous approach [7], we consider three typical learning applications:

  • (1)

    the basic task of learning a 3-dimensional rotation matrix

  • (2)

    learning the reachable space of a robot

  • (3)

    learning a point-to-point motion model from human demonstrations.

Following sections will present these three applications.

Discussion and conclusion

In this paper, we presented an adaption to standard GMM learning to accurately encode orientations with Euclidean space variables in a common model. We employed quaternion displacement, displacement integration and weighted averaging functions at all corresponding steps of the learning and regression algorithm. The validity of the approach was successfully shown on an artificial data set, learning a full turn of rotations about the z axis, as well as on two real-world data sets, learning the

Acknowledgment

This work was supported by the DFG Center of Excellence EXC 277: Cognitive Interaction Technology (CITEC).

Seungsu Kim received his Ph.D. in robotics from the Swiss Federal Institute of Technology in Lausanne (EPFL), Lausanne, Switzerland, in 2014. He is currently a Postdoctoral Fellow with the Neuroinformatics Group, CITEC, University of Bielefeld. He was a Researcher with the Center for Cognitive Robotics Research, Korea Institute of Science and Technology, Seoul, Korea, from 2005 to 2009. In 2015, he was awarded the IEEE Transactions on Robotics King-Sun Fu Memorial Best Paper Award. His research

References (29)

  • CalinonS. et al.

    On learning, representing, and generalizing a task in a humanoid robot

    IEEE Trans. Syst. Man Cybern. B

    (2007)
  • E. Gribovskaya, A. Billard, Learning nonlinear multi-variate motion dynamics for real-time position and orientation...
  • Khansari-ZadehS.M. et al.

    Learning stable nonlinear dynamical systems with gaussian mixture models

    IEEE Trans. Robot.

    (2011)
  • CalinonS. et al.

    Learning and reproduction of gestures by imitation

    IEEE Robot. Autom. Mag.

    (2010)
  • Khansari-ZadehS.M. et al.

    Learning to play minigolf: A dynamical system-based approach

    Adv. Robot.

    (2012)
  • S. Kim, E. Gribovskaya, A. Billard, Learning motion dynamics to catch a moving object, in: IEEE-RAS International...
  • KimS. et al.

    Catching objects in flight

    IEEE Trans. Robot.

    (2014)
  • FeitenW. et al.

    6D pose uncertainty in robotic perception

  • W. Feiten, M. Lang, MPG—A framework for reasoning on 6 dof pose uncertainty, in: ICRA Workshop on Manipulation under...
  • M. Lang, W. Feiten, MPG—Fast forward reasoning on 6 DOF pose uncertainty, in: Robotics; Proceedings of ROBOTIK 2012;...
  • DempsterA. et al.

    Maximum likelihood from incomplete data via the EM algorithm

    J. R. Stat. Soc. Ser. B Stat. Methodol.

    (1977)
  • BishopC.

    Pattern Recognition and Machine Learning

    (2006)
  • GörürD. et al.

    Dirichlet process Gaussian mixture models: Choice of the base distribution

    J. Comput. Sci. Tech.

    (2010)
  • TanakaK. et al.

    A quantum-statistical-mechanical extension of Gaussian mixture model

    J. Phys. Conf. Ser.

    (2008)
  • Cited by (36)

    View all citing articles on Scopus

    Seungsu Kim received his Ph.D. in robotics from the Swiss Federal Institute of Technology in Lausanne (EPFL), Lausanne, Switzerland, in 2014. He is currently a Postdoctoral Fellow with the Neuroinformatics Group, CITEC, University of Bielefeld. He was a Researcher with the Center for Cognitive Robotics Research, Korea Institute of Science and Technology, Seoul, Korea, from 2005 to 2009. In 2015, he was awarded the IEEE Transactions on Robotics King-Sun Fu Memorial Best Paper Award. His research interests include machine-learning techniques for robot manipulation.

    Robert Haschke received his Ph.D. in Computer Science from the University of Bielefeld, Germany, in 2004, working on the theoretical analysis of oscillating recurrent neural networks. He heads the Robotics Group within the Neuroinformatics Group, working on a bimanual robot setup for interactive learning. His fields of research include recurrent neural networks, cognitive bimanual robotics, grasping and manipulation with multifingered dexterous hands, tactile sensing, and software integration.

    Helge J. Ritter is the head of the Neuroinformatics Group at the Faculty of Technology, Bielefeld University. His main interests are principles of neural computation and intelligent systems, in particular cognitive robots with “manual intelligence”. In 1999, he was awarded the SEL Alcatel Research Prize and in 2001 the Leibniz Prize of the German Research Foundation DFG. He is a co-founder and Director of Bielefeld Cognitive Robotics Laboratory (CoR-Lab) and the coordinator of Bielefeld Excellence Cluster “Cognitive Interaction Technology” (CITEC).

    View full text