Nonlinear Synchronization for Automatic Learning of 3D Pose Variability in Human Motion Sequences
- 872 Downloads
A dense matching algorithm that solves the problem of synchronizing prerecorded human motion sequences, which show different speeds and accelerations, is proposed. The approach is based on minimization of MRF energy and solves the problem by using Dynamic Programming. Additionally, an optimal sequence is automatically selected from the input dataset to be a time-scale pattern for all other sequences. The paper utilizes an action specific model which automatically learns the variability of 3D human postures observed in a set of training sequences. The model is trained using the public CMU motion capture dataset for the walking action, and a mean walking performance is automatically learnt. Additionally, statistics about the observed variability of the postures and motion direction are also computed at each time step. The synchronized motion sequences are used to learn a model of human motion for action recognition and full-body tracking purposes.
KeywordsHuman Motion Training Sequence Dynamic Time Warping Human Posture Motion Capture Data
The nature of the open problems and techniques used in human motion analysis approaches strongly depends on the goal of the final application. Hence, most approaches oriented to surveillance demand performing activity recognition tasks in real-time dealing with illumination changes and low-resolution images. Thus, they require robust techniques with a low computational cost, and mostly, they tend to use simple models and fast algorithms to achieve effective segmentation and recognition tasks in real-time.
In contrast, approaches focused on 3D tracking and reconstruction require to deal with a more detailed representation about the current posture that the human body exhibits [4, 5, 6]. The aim of full body tracking is to recover the body motion parameters from image sequences dealing with 2D projection ambiguities, occlusion of body parts, and loose fitting clothes among others.
Many action recognition and 3D body tracking works rely on proper models of human motion, which constrain the search space using a training dataset of prerecorded motions [7, 8, 9, 10]. Consequently, it is highly desirable to extract useful information from the training set of motion. Traditional treatment suffers from problems inadequate modeling of nonlinear dynamics: training sequences may be acquired under very different conditions, showing different durations, velocities, and accelerations during the performance of an action. As a result, it is difficult to collect useful statistics from the raw training data, and a method for synchronizing the whole training set is required. Similarly to our work, in  a variation of DP is used to match motion sequences acquired from a motion capture system. However, the overall approach is aimed to the optimization of a posterior key-frame search algorithm. Then, the output from this process is used for synthesizing realistic human motion by blending the training set.
The DP approach has been widely used in literature for stereo matching and image processing applications [12, 13, 14]. Such applications often demand fast calculations in real-time, robustness against image discontinuities, and unambiguous matching.
The DP technique is a core of the dynamic time warping (DTW) method. Dynamic time warping is often used in speech recognition to determine if two waveforms represent the same spoken phrase . In addition to speech recognition, dynamic time warping has also been found useful in many other disciplines, including data mining, gesture recognition, robotics, manufacturing, and medicine .
Initially DTW method was developed for the one-dimensional signal processing (in speech recognition, e.g.). So, for this kind of the signal the Euclidean distance minimization with a weak constraint (the derivative of the synchronization path is constrained) works very well. In our case the dimensionality of the signal is up to 37D and weak constraint does not yield satisfactory robustness due to the noise and the signal complexity. We propose to minimize a composite distance that consists of two terms: a distance itself and a smoothness term. Such kind of a distance has the same meaning of the energy in MRF optimization techniques.
The MRF energy minimization approach shows the perfect performance in stereo matching and segmentation. Likewise, we present a dense matching algorithm based on DP, which is used to synchronize human motion sequences of the same action class in the presence of different speeds and accelerations. The algorithm finds an optimal solution in real-time.
We introduce a median sequences or the best pattern for time synchronization, which is another contribution of this work. The median sequence is automatically selected from the training data following a minimum global distance criterion among other candidates of the same class.
We present an action-specific model of human motion suitable for many applications, that has been successfully used for full body tracking [4, 5, 17]. In this paper, we explore and extend its capabilities for gait analysis and recognition tasks. Our action-specific model is trained with 3D motion capture data for the walking action from the CMU Graphics Lab Motion capture database. In our work, human postures are represented by means of a full body 3D model composed of 12 limbs. Limbs' orientations are represented within the kinematic tree using their direction cosines . As a result, we avoid singularities and abrupt changes due to the representation. Moreover, near configurations of the body limbs account for near positions in our representation at the expense of extra parameters to be included in the model. Then, PCA is applied to the training data to perform dimensionality reduction over the highly correlated input data. As a result, we obtain a lower-dimensional representation of human postures which is more suitable to describe human motion, since we found that each dimension on the PCA space describes a natural mode of variation of human motion. Additionally, the main modes of variation of human gait are naturally represented by means of the principal components found. This leads to a coarse-to-fine representation of human motion which relates the precision of the model with its complexity in a natural way and makes it suitable for different kinds of applications which demand more or less complexity in the model.
The synchronized version of the training set is utilized to learn an action-specific model of human motion. The observed variances from the synchronized postures of the training set are computed to determine which human postures can be feasible during the performance of a particular action. This knowledge is subsequently used in a particle filter tracking framework to prune those predictions which are not likely to be found in that action.
This paper is organized as follows. Section 2 explains the principles of human action modeling. In Section 3 we introduce a new dense matching algorithm for human motion sequences synchronization. Section 4 shows some examples of data base syncronisation. Section 5 describes the action specific model and explains the procedure for learning its parameters from the synchronized training set. Section 6 summarizes our conclusions.
2. Human Action Model
We use directional cosines to represent relative orientations of the limbs within the kinematic tree . As a result, we represent a human body posture Ψ using 37 parameters, that is,
where Open image in new window is the normalized height of the pelvis, and Open image in new window , Open image in new window , Open image in new window are the relative directional cosines for limb Open image in new window , that is, the cosine of the angle between a limb Open image in new window and each axis Open image in new window , Open image in new window , and Open image in new window , respectively. Directional cosines constitute a good representation method for body modeling, since it does not lead to discontinuities, in contrast to other methods such as Euler angles or spherical coordinates. Additionally, unlike quaternion, they have a direct geometric interpretation. However, given that we are using 3 parameters to determine only 2 DOFs for each limb, such representation generates a considerable redundancy of the vector space components. Therefore, we aim to find a more compact representation of the original data to avoid redundancy.
Let us introduce a particular performance of an action. A performance Open image in new window consists of a time-ordered sequence of postures
where Open image in new window is an index indicating the number of performance, andF i is the total number of postures that constitute the performance Open image in new window . We assume that each two consecutive postures are separated by a time interval δ f, which depends on the frame rate of the prerecorded input sequences; thus the duration of a particular performance is Open image in new window . Finally, an action A k is defined by all the I k performances that belong to that action Open image in new window
As we mentioned above, the original vector space is redundant. Additionally, the human body motion is intrinsically constrained, and these natural constraints lead to highly correlated data in the original space. Therefore, we aim to find a more compact representation of the original data to avoid redundancy. To do this, we consider a set of performances corresponding to a particular action A k and perform the Principal Component Analysis (PCA) to all the postures that belong to that action. Eventually, the following eigenvector decomposition equation has to be solved:
where Open image in new window stands for the Open image in new window covariance matrix calculated with all the postures of action A k . As a result, each eigenvector Open image in new window corresponds to a mode of variation of human motion, and its corresponding eigenvalue λ j is related to the variance specified by the eigenvector. In our case, each eigenvector reflects a natural mode of variation of human gait. To perform dimensionality reduction over the original data, we consider only the first b eigenvectors that span the new representation space for this action, hereafter aSpace. We assume that the overall variance of a new space approximately equals to the overall variance of the unreduced space:
where ε b is the aSpace approximation error.
Consequently, we use (4) to find the smallest number b of eigenvalues, which provide an appropriate approximation of the original data, and human postures are projected into the aSpace by
where Open image in new window refers to the original posture, Open image in new window denotes the lower-dimensional version of the posture represented using the aSpace, [e1,…,e b ] is the aSpace transformation matrix that correspond to the first b selected eigenvectors, and Open image in new window is the posture mean value that is formed by averaging all postures, which are assumed to be transformed into the aSpace. As a result, we obtain a lower-dimensional representation of human postures which is more suitable to describe human motion, since we found that each dimension on the PCA space describes a natural mode of variation of human motion . Choosing different values for b lead to models of more or less complexity in terms of their dimensionality. Hence, while the gross-motion(mainly, the motion of the torso, legs, and arms in low resolution) is explained by the very first eigenvectors, subtle motions in the PCA space representation require more eigenvectors to be considered. In other words, the initial 37-dimensional parametric space becomes the restricted b-dimensional parametric space.
The projection of the training sequences into the aSpace constitutes the input for our sequence synchronization algorithm. Hereafter, we consider a multidimensional signal x i (t) as an interpolated expansion of each training sequence Open image in new window such as
where the time domain of each action performance x i (t) is [0,Ti).
3. Synchronization Algorithm
As stated before, the training sequences are acquired under very different conditions, showing different durations, velocities, and accelerations during the performance of a particular action. As a result, it is difficult to perform useful statistical analysis to the raw training set, since we cannot put in correspondence postures from different cycles of the same action. Therefore, a method for synchronizing the whole training set is required so that we can establish a mapping between postures from different cycles.
where x n and x m are the two compared multidimensional signals, T n and T m are the periods of the action performances n and m, and Open image in new window is linearly normalized version of Open image in new window ; hence Open image in new window .
Unfortunately, in our research we rarely, if ever, have a constant rate ratio α. An example, which is illustrated in Figure 2(b), shows that a simple normalization using (7) does not give us the needed signal fitting, and a nonlinear data synchronization method is needed. Further in the text we will assume that the linear synchronization is done and all the periods T n possess the same value T.
The nonlinear data synchronization should be done by
where x n,m (t) is the best synchronized version of the action x m (t) to the action x n (t). In literature the function τ(t) is usually referred to as the distance-time function. It is not an apt turn of phrase indeed, and we suggest naming it as the rate-to-rate synchronization function instead.
The rate-to-rate synchronization function τ(t) satisfies several useful constraints, that are
One common approach for building the function τ(t) is based on a key-frame model. This model assumes that the compared signals x n and Open image in new window have similar sets of singular points, that are Open image in new window and Open image in new window with the matching condition Open image in new window . The aim is to detect and match these singular points; thus the signals x n and x m are synchronized. However, the singularity detection is an intricate problem itself, and to avoid the singularity detection stage we propose a dense matching. In this case a time interval Open image in new window is constant, and in general Open image in new window .
Let us introduce a formal measure of synchronization of two signals by
where Open image in new window denotes one of possible vector distances, and D n,m is referred to as the synchronization distance that consists of two parts, where the first integral represents the functional distance between the two signals, and the second integral is a regularization term, which expresses desirable smoothness constraints of the solution. The proposed distance function is simple and makes intuitive sense. It is natural to assume that the compared signals are synchronized better when the synchronization distance between them is minimal. Thus, the sought function Δ n,m (t) should minimize the synchronization distance between matched signals.
In the case of a discrete time representation, (11) can be rewritten as
where δ t is a time sampling interval. Equation (9) implies
The synchronization problem is similar to the matching problem of two epipolar lines in a stereo image. In the case of the stereo image processing the parameter Δ(t) is called disparity. For stereo matching a DSI representation is used. The DSI approach assumes that 2D DSI matrix has dimensions time Open image in new window , and disparity Open image in new window . Let Open image in new window denote the DSI cost value assigned to matrix element Open image in new window and calculated by
Now we formulate an optimization problem as follows: find the time-disparity function Δ n,m (p), which minimizes the synchronization distance between the compared signals x n and x m , that is,
The optimal path problem can be easily solved by using the method of dynamic programming. The method consists of step-by-step control and optimization that is given by a recurrence relation:
where the scope of the minimization parameter Open image in new window is chosen in accordance with (13). By using the recurrence relation the minimal value of the objective function in (15) can be found at the last step of optimization. Next, the algorithm works in reverse order and recovers a sequence of optimal steps (using the lookup table K(p,d) of the stored values of the index k in the recurrence relation (16)) and eventually the optimal path by
Now the synchronized version of x m (t) might be easily calculated by
Here we assume that n is the number of the base rate sequences and m is the number of sequences to be synchronized.
Our algorithm assumes that a particular sequence is chosen to be a time scale pattern for all other sequences. It is obvious that an arbitrary choice among the training set is not a reasonable solution, and now we aim to find a statistically proven rule that is able to make an optimal choice according to some appropriate criterion. Note that each synchronized pair of sequences (n,m) has its own synchronization distance calculated by (12). Then the full synchronization of all the sequences relative to the pattern sequences n has its own global distance:
We propose to choose the synchronizing pattern sequence with minimal global distance. In statistical sense such signal can be considered as a median value over all the performances that belong to the set of Open image in new window or can be referred to as "median'' sequence.
4. Results of Synchronization
To prove the correctness of our approach, we manually synchronized the same training set by selecting a set of 5 key-frames in each sequence by hand following a maximum curvature subjective criterion. Then, the training set was resampled; so each sequence had the same number of frames between each key-frame. In Figure 5(c), the first 4 dimensions within the aSpace of the resulting manually synchronized sequences are shown. We might observe that the results are very similar to the ones obtained with the proposed automatic synchronization method. The synchronized training set from Figure 5(b) has been used to learn an action-specific model of human motion for the bending action. The model learns a mean-performance for the synchronized training set and its observed variance at each posture. In Figure 5(d) the learnt action model for the bending action is plotted. The mean-performance corresponds to the solid red line while the black solid line depicts ±3 times the learnt standard deviation at each synchronized posture. The input training sequence set is depicted as dashed blue lines.
Additionally, to prove the advantage of our approach with respect to DTW we applied our algorithm with the cut objective function (without smoothness term), which is coincide with the DTW algorithm. In this case the synchronization process was not satisfactory: some selected mean postures were completely outliers or nonsimilar to any meaningful posture. It means that the smoothness factor Open image in new window in (12) and (16) plays an important role. To find an optimal value of this parameter a visual criterion has been used (the manual synchronization that had been done before yields such a visual estimation technique). However, as a rule of thumb the parameter can be set equal to the mean value of the error term E(i,d):
5. Learning the Motion Model
Once all the sequences share the same time pattern, we learn an action specific model which is accurate without loosing generality and suitable for many applications. In this section we consider the waking action and its model is useful for gait analysis, gait recognition, and tracking. Thus, we want to learn where the postures lie in the space used for representation, how they change over time as the action goes by, and what characteristics the different performances have in common which can be exploited for enabling the aforementioned tasks. In other words, we aim to characterize the shape of the synchronized version of the training set for the walking action in the PCA-like space. The process is as follows.
First, we extract from the training set Open image in new window a mean representation of the action by computing the mean performance Open image in new window where each mean posture Open image in new window is defined as
Open image in new window is the number of training performances for the action Open image in new window , Open image in new window corresponds to the Open image in new window th posture from the Open image in new window th training performance, and finally, Open image in new window denotes the total number of postures of each synchronized performance.
Then, we want to quantify how much the training performances Open image in new window vary from the computed mean performance Open image in new window of (21). Therefore, for each time step t, we compute the standard deviation Open image in new window of all the postures Open image in new window that share the same time stamp t, that is,
On the other hand, we are also interested in characterizing the temporal evolution of the action. Therefore, we compute the main direction of the motion Open image in new window for each subsequence of Open image in new window postures from the mean performance Open image in new window , that is,
Each black arrow corresponds to the unitary vector Open image in new window computed at time t, scaled for visualization purposes. Hence, each vector encodes the mean observed motion's direction from time Open image in new window to time t, where d stands for the length of the motion window considered. Additionally, selected postures from the mean performance have been sampled at times Open image in new window , and 168 and overlaid in the graphic.
As a result, the action model Open image in new window is defined by
where Open image in new window is the PCA space definition for action Open image in new window , Open image in new window is the mean performance, and Open image in new window and Open image in new window correspond to the computed standard deviation and mean direction of motion at each time step t, respectively.
Finally, to handle the cyclic nature of the waking action, we concatenate the last postures in each cycle with the initial postures of the most close performance according to a Euclidean distance criterion within the PCA space. Additionally, the first and last d/2 postures from the mean performance (where d is the length of the considered subsequences) are resampled using cubic spline interpolation in order to soft the transition between walking cycles. As a result, we are able to compute Open image in new window , Open image in new window for the last postures of a full walking cycle.
6. Conclusions and Future Work
In this paper, a novel dense matching algorithm for human motion sequences synchronization has been proposed. The technique utilizes dynamic programming and can be used in real-time applications. We also introduce the definition of the median sequence that is used to choose a time-scale pattern for all other sequences. The synchronized motion sequences are utilized to learn a model of human motion and to extract signal statistics. We have presented an action-specific model suitable for gait analysis, gait identification and tracking applications. The model is tested for the walking action and is automatically learnt from the public CMU motion capture database. As a result, we learnt the parameters of our action model which characterize the pose variability observed within a set of walking performances used for training.
The resulting action model consists of a representative manifold for the action, namely, the mean performance, the standard deviation from the mean performance. The action model can be used to classify which postures belong to the action or not. Moreover, the tradeoff between accuracy and generality of the model can be tuned using more or less dimensions for building the PCA space representation of human postures. Hence, using this coarse-to-fine representation, the main modes of variation correspond to meaningful natural motion modes. Thus, for example, we found that the main modes of variation for the walking action obtained from PCA explain the combined motion of both the legs and the arms, while in the bending action they mainly correspond to the motion of the torso.
Future research lines rely on obtaining the joint positions directly from image sequences. Previously, the action model has been successfully used in a probabilistic tracking framework for estimating the parameters of our 3D model from a sequence of 2D images. In  the action model improved the efficiency of the tracking algorithm by constraining the space of possible solutions only to the most feasible postures while performing a particular action, thus avoiding estimating postures which are not likely to occur during an action. However, we need to develop robust image-based likelihood measures which evaluate the predictions from our action model according to the measurements obtained from images. Work based on extracting the image edges and the silhouette from the tracked subject is currently in progress. Hence, the pursued objective is to learn a piecewise linear model which evaluates the fitness of segmented edges and silhouettes to the 2D projection of the stick figure from our human body model. Methods for estimating the 6DOF of the human body within the scene, namely, 3D translation and orientation, also need to be improved.
This work has been supported by EC Grant IST-027110 for the HERMES project and by the Spanish MEC under projects TIC2003-08865 and DPI-2004-5414. M. Mozerov acknowledges the support of the Ramon y Cajal research program, MEC, Spain.
- 6.Sigal L, Black MJ: Measure locally, reason globally: occlusion-sensitive articulated pose estimation. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR '06), 2006 2: 2041-2048.Google Scholar
- 8.Roberts TJ, McKenna SJ, Ricketts IW: Adaptive learning of statistical appearance models for 3D human tracking. Proceedings of the British Machine Vision Conference (BMVC '02), September 2002, Cardiff, UK 121-165.Google Scholar
- 9.Sidenbladh H, Black MJ, Sigal L: Implicit probabilistic models of human motion for synthesis and tracking. In Proceedings of the 7th European Conference on Computer Vision Copenhagen (ECCV '02), May 2002, Copenhagen, Denmark, Lecture Notes in Computer Science. Volume 2350. Springer; 784-800.Google Scholar
- 11.Nakazawa A, Nakaoka S, Ikeuchi K: Matching and blending human motions using temporal scaleable dynamic programming. Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS '04), September-October 2004, Sendai, Japan 1: 287-294.Google Scholar
- 15.Kruskall J, Liberman M: The symmetric time warping problem: from continuous to discrete. In Time Warps, String Edits, and Macromolecules: The Theory and Practice of Sequence Comparison. Madison-Wesley, Reading, Mass, USA; 1983:125-161.Google Scholar
- 16.Keogh E, Pazzani M: Derivative dynamic time warping. Proceedings of the 1st SIAM International Conference on Data Mining, 2001, Chicago, Ill, USA 1-12.Google Scholar
- 17.Rius I, Rowe D, Gonzalez J, Xavier Roca F: 3D action modeling and reconstruction for 2D human body tracking. Proceedings of the 3rd International Conference on Advances in Patten Recognition (ICAPR '05), August 2005, Bath, UK 3687: 146-154.Google Scholar
- 18.Zatsiorsky VM: Kinematics of Human Motion. Human Kinematics; 1998.Google Scholar
This article is published under license to BioMed Central Ltd. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.