Cueing, feature discovery, and one-class learning for synthetic aperture radar automatic target recognition

The exquisite capabilities of biological neural systems for recognizing target patterns subject to large variations have motivated us to investigate neurophysiologically-inspired techniques for automatic target recognition. This paper describes a modular multi-stage architecture for focus-of-attenti...

Full description

Saved in:
Bibliographic Details
Published inNeural networks Vol. 8; no. 7; pp. 1081 - 1102
Main Authors Koch, Mark W., Moya, Mary M., Hostetler, Larry D., Fogler, R.Joseph
Format Journal Article
LanguageEnglish
Published Oxford Elsevier Ltd 01.01.1995
Elsevier Science
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:The exquisite capabilities of biological neural systems for recognizing target patterns subject to large variations have motivated us to investigate neurophysiologically-inspired techniques for automatic target recognition. This paper describes a modular multi-stage architecture for focus-of-attention cueing, feature discovery and extraction, and one-class pattern learning and identification in synthetic aperture radar imagery. To prescreen massive amounts of image data, we apply a focus-of-attention algorithm using data skewness to extract man-made objects from natural clutter regions. We apply self-organizing feature discovery algorithms that uniquely characterize targets in a reduced dimension space and use self-organizing one-class classifiers for learning target variations. We also develop a distance metric for partial obscuration recognition. We present performance results using simulated SAR data and test for within-class generalization using nontrained targets including both in-the-clear and partially obscured examples. We test for between-class generalization using non-trained targets including both in-the-clear and partially obscured examples. We test for between-class generalization using near-target data.
Bibliography:ObjectType-Article-2
SourceType-Scholarly Journals-1
ObjectType-Feature-1
content type line 23
ISSN:0893-6080
1879-2782
DOI:10.1016/0893-6080(95)00049-6