A spatio-temporal representation for complex optical flow events is developed that generalizes traditional parameterized motion models (e.g. affine). These generative spatio-temporal models may be non-linear or stochastic and are event-specific in that they characterize a particular type of object motion (e.g. sitting or walking). Within a Bayesian framework we seek the appropriate model, phase, rate, spatial position, and scale to account for the image variation. The posterior distribution over this parameter space conditioned on image measurements is typically nonGaussian. The distribution is represented using factored sampling and is predicted and updated over time using the Condensation algorithm. The resulting framework automatically detects, localizes, and recognizes motion events.
Michael J. Black