Modeling the Distribution of Normal Data in Pre-Trained Deep Features for Anomaly Detection

Anomaly Detection (AD) in images is a fundamental computer vision problem and refers to identifying images and/or image substructures that deviate significantly from the norm. Popular AD algorithms commonly try to learn a model of normality from scratch using task specific datasets, but are limited...

Full description

Saved in:
Bibliographic Details
Published in2020 25th International Conference on Pattern Recognition (ICPR) pp. 6726 - 6733
Main Authors Rippel, Oliver, Mertens, Patrick, Merhof, Dorit
Format Conference Proceeding
LanguageEnglish
Published IEEE 10.01.2021
Subjects
Online AccessGet full text

Cover

Loading…
Abstract Anomaly Detection (AD) in images is a fundamental computer vision problem and refers to identifying images and/or image substructures that deviate significantly from the norm. Popular AD algorithms commonly try to learn a model of normality from scratch using task specific datasets, but are limited to semi-supervised approaches employing mostly normal data due to the inaccessibility of anomalies on a large scale combined with the ambiguous nature of anomaly appearance. We follow an alternative approach and demonstrate that deep feature representations learned by discriminative models on large natural image datasets are well suited to describe normality and detect even subtle anomalies in a transfer learning setting. Our model of normality is established by fitting a multivariate Gaussian (MVG) to deep feature representations of classification networks trained on ImageNet using normal data only. By subsequently applying the Mahalanobis distance as the anomaly score we outperform the current state of the art on the public MVTec AD dataset, achieving an Area Under the Receiver Operating Characteristic curve of 95.8 ± 1.2% (mean ± SEM) over all 15 classes. We further investigate why the learned representations are discriminative to the AD task using Principal Component Analysis. We find that the principal components containing little variance in normal data are the ones crucial for discriminating between normal and anomalous instances. This gives a possible explanation to the often subpar performance of AD approaches trained from scratch using normal data only. By selectively fitting a MVG to these most relevant components only, we are able to further reduce model complexity while retaining AD performance. We also investigate setting the working point by selecting acceptable False Positive Rate thresholds based on the MVG assumption. Code is publicly available at https://github.com/ORippler/gaussian-ad-mvtec.
AbstractList Anomaly Detection (AD) in images is a fundamental computer vision problem and refers to identifying images and/or image substructures that deviate significantly from the norm. Popular AD algorithms commonly try to learn a model of normality from scratch using task specific datasets, but are limited to semi-supervised approaches employing mostly normal data due to the inaccessibility of anomalies on a large scale combined with the ambiguous nature of anomaly appearance. We follow an alternative approach and demonstrate that deep feature representations learned by discriminative models on large natural image datasets are well suited to describe normality and detect even subtle anomalies in a transfer learning setting. Our model of normality is established by fitting a multivariate Gaussian (MVG) to deep feature representations of classification networks trained on ImageNet using normal data only. By subsequently applying the Mahalanobis distance as the anomaly score we outperform the current state of the art on the public MVTec AD dataset, achieving an Area Under the Receiver Operating Characteristic curve of 95.8 ± 1.2% (mean ± SEM) over all 15 classes. We further investigate why the learned representations are discriminative to the AD task using Principal Component Analysis. We find that the principal components containing little variance in normal data are the ones crucial for discriminating between normal and anomalous instances. This gives a possible explanation to the often subpar performance of AD approaches trained from scratch using normal data only. By selectively fitting a MVG to these most relevant components only, we are able to further reduce model complexity while retaining AD performance. We also investigate setting the working point by selecting acceptable False Positive Rate thresholds based on the MVG assumption. Code is publicly available at https://github.com/ORippler/gaussian-ad-mvtec.
Author Merhof, Dorit
Rippel, Oliver
Mertens, Patrick
Author_xml – sequence: 1
  givenname: Oliver
  surname: Rippel
  fullname: Rippel, Oliver
  email: oliver.rippel@lfb.rwth-aachen.de
  organization: Institute of Imaging & Computer Vision, RWTH Aachen University,Aachen,Germany
– sequence: 2
  givenname: Patrick
  surname: Mertens
  fullname: Mertens, Patrick
  organization: Institute of Imaging & Computer Vision, RWTH Aachen University,Aachen,Germany
– sequence: 3
  givenname: Dorit
  surname: Merhof
  fullname: Merhof, Dorit
  organization: Institute of Imaging & Computer Vision, RWTH Aachen University,Aachen,Germany
BookMark eNotj81KAzEcxCPowVafQJC8wK752mxyLLvWFqoW7c1DyWb_0cA2Kdn00Ld3xZ4GZoYfMzN0HWIAhB4pKSkl-mndbD-EUkSWjDBaakHZZF-hGa2ZolOg-C36eo09DD584_wDuPVjTr47ZR8Djg6_xXQwA25NNtgHvE1Q7JLxAXrcAhzxEkw-JRixiwkvQpzK5ynJYP8Id-jGmWGE-4vO0efyedesis37y7pZbArPpM4F04LwzkoQSndK9rzi0zRqqHLWKldXQjpttBQAujK2Mh2rjaZOWyJ0x-fo4Z_qAWB_TP5g0nl_ect_AXUET3Q
ContentType Conference Proceeding
DBID 6IE
6IL
CBEJK
RIE
RIL
DOI 10.1109/ICPR48806.2021.9412109
DatabaseName IEEE Electronic Library (IEL) Conference Proceedings
IEEE Xplore POP ALL
IEEE Xplore All Conference Proceedings
IEEE Electronic Library (IEL)
IEEE Proceedings Order Plans (POP All) 1998-Present
DatabaseTitleList
Database_xml – sequence: 1
  dbid: RIE
  name: IEEE Electronic Library (IEL)
  url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/
  sourceTypes: Publisher
DeliveryMethod fulltext_linktorsrc
EISBN 1728188083
9781728188089
EndPage 6733
ExternalDocumentID 9412109
Genre orig-research
GroupedDBID 6IE
6IL
CBEJK
RIE
RIL
ID FETCH-LOGICAL-i269t-29403bc6e489b86d3530831a18fcc8f7546f9a964ee95ac5ab27a91f9c049b3
IEDL.DBID RIE
IngestDate Wed Aug 27 02:29:58 EDT 2025
IsPeerReviewed false
IsScholarly true
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-i269t-29403bc6e489b86d3530831a18fcc8f7546f9a964ee95ac5ab27a91f9c049b3
PageCount 8
ParticipantIDs ieee_primary_9412109
PublicationCentury 2000
PublicationDate 2021-01-10
PublicationDateYYYYMMDD 2021-01-10
PublicationDate_xml – month: 01
  year: 2021
  text: 2021-01-10
  day: 10
PublicationDecade 2020
PublicationTitle 2020 25th International Conference on Pattern Recognition (ICPR)
PublicationTitleAbbrev ICPR
PublicationYear 2021
Publisher IEEE
Publisher_xml – name: IEEE
Score 2.5834029
Snippet Anomaly Detection (AD) in images is a fundamental computer vision problem and refers to identifying images and/or image substructures that deviate...
SourceID ieee
SourceType Publisher
StartPage 6726
SubjectTerms Computer vision
Data models
Feature extraction
Fitting
Receivers
Training
Transfer learning
Title Modeling the Distribution of Normal Data in Pre-Trained Deep Features for Anomaly Detection
URI https://ieeexplore.ieee.org/document/9412109
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3PS8MwFA7bTp5UNvE3OXi0XX-mzVFWxxQ2hk4YeBgv6QsMpR3SHvSvN6-rE8WDt5ImbUgC33sv73sfY1dhqFFAohwFQULRqsSRAsDxNXgReMZCEsUhpzMxeYrul_Gyw653XBhEbJLP0KXH5i4_L3VNobKhjKjcleyyrnXctlytlvRrW4d3o_kDHUdKPAh8t-38QzWlAY3xPpt-_W6bK_Li1pVy9cevSoz_nc8BG3zT8_h8BzyHrINFnz2Tqhlxy7k16XhG9XBbKSteGj4j0_SVZ1ABXxd2NDoLEofAnGeIG06WYG09b25tWH5TlLbzu31TNYlaxYA9jm8Xo4nTKic460DIyglk5IVKC4xSqVKRh3FIimLgp0br1CRxJIwEKSJEGYOOQQUJSN9IbR0GFR6xXlEWeMx4LkWqEzQ614qudCEW2qhUg_2-BXZ1wvq0LKvNtjTGql2R07-bz9gebQ1FMHzvnPWqtxovLKZX6rLZzE-5YaNw
linkProvider IEEE
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3PS8MwFA5zHvSksom_zcGj7foraXOUzbHpNoZOGHgYSfoCQ2mHtAf9683r6kTx4C0kr21ICt_Ly_veR8hVGGrgMlaOkkGM0arYEVxKx9fSi6RnLCRhHHI84YOn6G7O5g1yveHCAECVfAYuNqu7_DTXJYbKOiLCcldii2xb3GfBmq1V035tf2fYnT7gD4mpB4Hv1uY_dFMq2OjvkfHXB9fZIi9uWShXf_yqxfjfGe2T9jdBj0430HNAGpC1yDPqmiG7nFqnjvawIm4tZkVzQyfonL7SniwkXWb2aXBmKA8BKe0BrCj6gqU9e1PrxdKbLLfG73akqFK1sjZ57N_OugOn1k5wlgEXhROIyAuV5hAlQiU8DVmImmLST4zWiYlZxI2QgkcAgknNpApiKXwjtD0yqPCQNLM8gyNCU8ETHYPRqVZ4qSsZ10YlWtr3W2hXx6SFy7JYrYtjLOoVOfm7-5LsDGbj0WI0nNyfkl3cJoxn-N4ZaRZvJZxbhC_URbWxn3GRpro
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=2020+25th+International+Conference+on+Pattern+Recognition+%28ICPR%29&rft.atitle=Modeling+the+Distribution+of+Normal+Data+in+Pre-Trained+Deep+Features+for+Anomaly+Detection&rft.au=Rippel%2C+Oliver&rft.au=Mertens%2C+Patrick&rft.au=Merhof%2C+Dorit&rft.date=2021-01-10&rft.pub=IEEE&rft.spage=6726&rft.epage=6733&rft_id=info:doi/10.1109%2FICPR48806.2021.9412109&rft.externalDocID=9412109