TR2010-071

Tracking Motion, Deformation, and Texture Using Conditionally Gaussian Processes


    •  Marks, T.K., Hershey, J.R., Movellan, J.R., "Tracking Motion, Deformation, and Texture Using Conditionally Gaussian Processes", IEEE Transactions on Pattern Analysis and Machine Intelligence, DOI: 10.1109/​TPAMI.2008.278, Vol. 32, No. 2, pp. 348-363, February 2010.
      BibTeX TR2010-071 PDF
      • @article{Marks2010feb,
      • author = {Marks, T.K. and Hershey, J.R. and Movellan, J.R.},
      • title = {Tracking Motion, Deformation, and Texture Using Conditionally Gaussian Processes},
      • journal = {IEEE Transactions on Pattern Analysis and Machine Intelligence},
      • year = 2010,
      • volume = 32,
      • number = 2,
      • pages = {348--363},
      • month = feb,
      • doi = {10.1109/TPAMI.2008.278},
      • url = {https://www.merl.com/publications/TR2010-071}
      • }
  • MERL Contact:
  • Research Area:

    Computer Vision

Abstract:

We present a generative model and inference algorithm for 3D nonrigid object tracking. The model, which we call G-flow, enables the joint inference of 3D position, orientation, and nonrigid deformations, as well as object texture and background texture. Optimal inference under G-flow reduces to a conditionally Gaussian stochastic filtering problem. The optimal solution to this problem reveals a new space of computer vision algorithms, of which classic approaches such as optic flow and template matching are special cases that are optimal only under special circumstances. We evaluate G-flow on the problem of tracking facial expressions and head motion in 3D from single-camera video. Previously, the lack of realistic video data with ground truth nonrigid position information has hampered the rigorous evaluation of nonrigid tracking. We introduce a practical method of obtaining such ground truth data and present a new face video data set that was created using this technique. Results on this data set show that G-flow is much more robust and accurate than current deterministic optic-flow-based approaches.

 

  • Related News & Events