Metrics for evaluating the performance of machine learning based automated valuation models

Automated Valuation Models (AVMs) based on Machine Learning (ML) algorithms are widely used for predicting house prices. While there is consensus in the literature that cross-validation (CV) should be used for model selection in this context, the interdisciplinary nature of the subject has made it h...

Full description

Saved in:
Bibliographic Details
Published inJournal of property research Vol. 38; no. 2; pp. 99 - 129
Main Authors Steurer, Miriam, Hill, Robert J., Pfeifer, Norbert
Format Journal Article
LanguageEnglish
Published Abingdon Routledge 03.04.2021
Taylor & Francis Ltd
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Automated Valuation Models (AVMs) based on Machine Learning (ML) algorithms are widely used for predicting house prices. While there is consensus in the literature that cross-validation (CV) should be used for model selection in this context, the interdisciplinary nature of the subject has made it hard to reach consensus over which metrics to use at each stage of the CV exercise. We collect 48 metrics (from the AVM literature and elsewhere) and classify them into seven groups according to their structure. Each of these groups focuses on a particular aspect of the error distribution. Depending on the type of data and the purpose of the AVM, the needs of users may be met by some classes, but not by others. In addition, we show in an empirical application how the choice of metric can influence the choice of model, by applying each metric to evaluate five commonly used AVM models. Finally - since it is not always practicable to produce 48 different performance metrics - we provide a short list of 7 metrics that are well suited to evaluate AVMs. These metrics satisfy a symmetry condition that we find is important for AVM performance, and can provide a good overall model performance ranking.
ISSN:0959-9916
1466-4453
DOI:10.1080/09599916.2020.1858937