Real-time Body Tracking Using a Gaussian Process Latent Variable Model


Abstract

In this paper, we present a tracking framework for capturing articulated human motions in real-time, without the need for attaching markers onto the subject's body. This is achieved by first obtaining a low dimensional representation of the training motion data, using a nonlinear dimensionality reduction technique called back-constrained GPLVM. A prior dynamics model is then learnt from this low dimensional representation by partitioning the motion sequences into elementary movements using an unsupervised EM clustering algorithm. The temporal dependencies between these elementary movements are efficiently captured by a Variable Length Markov Model. The learnt dynamics model is used to bias the propagation of candidate pose feature vectors in the low dimensional space. By combining this with an efficient volumetric reconstruction algorithm, our framework can quickly evaluate each candidate pose against image evidence captured from multiple views. We present results that show our system can accurately track complex structured activities such as ballet dancing in real-time.

For more details see:

  • S. Hou, A. Galata, N. Thacker and P. Bromiley, "Real-Time Body Tracking Using a Gaussian Process Latent Variable Model", in IEEE International Conference on Computer Vision (ICCV), 2007 ( .pdf ).

  • Demos

    Figure 1: Tracking results - Ballet dance exercise 1 (movie)



    Figure 2: Tracking results - Ballet dance exercise 2 (movie)

    Other Related Papers:

  • Fabrice Caillette, Aphrodite Galata, Toby Howard, "Real-Time 3-D Human Body Tracking using Learnt Models of Behaviour", Computer Vision and Image Understanding (CVIU) Journal, Vol. 109, No. 2, pp. 112 - 125, 2008. ( .pdf )

  • Fabrice Caillette, Aphrodite Galata, Toby Howard, "Real-Time 3-D Human Body Tracking using Variable Length Markov Models", In British Machine Vision Conference (BMVC), pp. 469-478, 2005. ( .pdf )

  • Back to Aphrodite's Home Page