Propagating covariance in computer vision

This paper describes how to propagate approximately additive random perturbations through any kind of vision algorithm step in which the appropriate random perturbation model for the estimated quantity produced by the vision step is also an additive random perturbation. The author assumes that the v...

Full description

Saved in:
Bibliographic Details
Published inPattern Recognition, 1994 12th International Conference On. Vol. 1 Vol. 1; pp. 493 - 498 vol.1
Main Author Haralick, R.M.
Format Conference Proceeding
LanguageEnglish
Published IEEE 1994
Subjects
Online AccessGet full text
ISBN0818662654
9780818662652
DOI10.1109/ICPR.1994.576335

Cover

Loading…
More Information
Summary:This paper describes how to propagate approximately additive random perturbations through any kind of vision algorithm step in which the appropriate random perturbation model for the estimated quantity produced by the vision step is also an additive random perturbation. The author assumes that the vision algorithm step can be modeled as a calculation (linear or nonlinear) that produces an estimate that minimizes an implicit scaler function of the input quantity and the calculated estimate. The only assumption is that the scaler functions have finite second partial derivatives and that the random perturbations are small enough so that the relationship between the scaler function evaluated at the ideal but unknown input and output quantities and the observed input quantity and perturbed output quantity can be approximated sufficiently well by a first order Taylor series expansion. The paper finally discusses the issues of verifying that the derived statistical behavior agrees with the experimentally observed statistical behavior.
ISBN:0818662654
9780818662652
DOI:10.1109/ICPR.1994.576335