10,580
Views
12
CrossRef citations to date
0
Altmetric
Article

Metrics for evaluating the performance of machine learning based automated valuation models

ORCID Icon, &
Pages 99-129 | Received 18 Jan 2020, Accepted 30 Nov 2020, Published online: 17 Apr 2021

ABSTRACT

Automated Valuation Models (AVMs) based on Machine Learning (ML) algorithms are widely used for predicting house prices. While there is consensus in the literature that cross-validation (CV) should be used for model selection in this context, the interdisciplinary nature of the subject has made it hard to reach consensus over which metrics to use at each stage of the CV exercise. We collect 48 metrics (from the AVM literature and elsewhere) and classify them into seven groups according to their structure. Each of these groups focuses on a particular aspect of the error distribution. Depending on the type of data and the purpose of the AVM, the needs of users may be met by some classes, but not by others. In addition, we show in an empirical application how the choice of metric can influence the choice of model, by applying each metric to evaluate five commonly used AVM models. Finally – since it is not always practicable to produce 48 different performance metrics – we provide a short list of 7 metrics that are well suited to evaluate AVMs. These metrics satisfy a symmetry condition that we find is important for AVM performance, and can provide a good overall model performance ranking.

1. Introduction

While parametric models remain the gold standard when it comes to understanding the structure of the world around us, data-driven semi- or non-parametric models – collectively often referred to as Machine Learning (ML) models – generally outperform their parametric counterparts at short-term out-of-sample prediction.Footnote1 Driven by the development of new methods, increased computing power, and the emergence of big data, the last two decades have brought about a large growth in new ML methods. A distinction can be drawn between those ML methods that predict numerical values and those that classify observations into different groups. Our focus here is the former task – in particular, we are interested in how researchers can judge the relative performance of competing Automated Valuation Models (AVMs) of the real estate market.

The objective of a housing AVM is to predict the price of apartments or houses.Footnote2 The traditional benchmark for AVMs is the hedonic model, where price (or log price) is assumed to depend on available characteristics in an additive way. This model has a number of benefits: it can be easily estimated via least-squares regression, it is well grounded in economic theory, and its output has an intuitive interpretation (total price is dependent on the shadow prices of the individual characteristics). However, due to their superior performance with regard to price prediction, most AVMs use some type of ML technique instead of hedonic models (see e.g. Schulz et al., Citation2014). When it comes to ML methods, users can choose from a large and continually expanding list of different approaches such as Random Forests, Quantile Regression, LASSO Regression, Adaptive Regression Splines, and Neural Nets, to name but a few.

Cross-Validation (CV) is a well-known model selection technique that can be used to compare the performance of parametric, non-parametric, and semi-parametric models. It provides a system for judging the performance of models on independent test samples and is the most popular model selection technique for ML methods (Yang, Citation2007). We use it here to compare the performance of the different ML models.

Two decisions need to be made to successfully use CV to judge model performance: first, how to organise the train/test-set split, and second, which metrics to use to judge performance. In this paper we focus on the role of performance metrics. We present 48 metrics that could potentially be used for this task. The interdisciplinary nature of the subject has made it hard for any consensus to emerge over the properties of these performance metrics and their suitability in particular contexts. We classify metrics into seven classes and in the process rationalise the relevant literature.

The paper is organised as follows: We begin with a general discussion on AVMs and metrics in section 2. In section 3 we survey a range of metrics that have been proposed in different strands of the literature and bring them together using a common notation to allow direct comparison. These metrics are classified into seven classes based on their structure. In total, we consider 48 different metrics that could be used to evaluate model performance. Although the focus of our analysis is on the housing market, the metrics discussed in this paper should be useful in all situations where a choice between different regression models needs to be made. In section 4 we discuss our dataset, cleaning procedure, outlier detection, variable creation and transformation, the construction of the CV folds (the train/test split), and model tuning and selection.

In section 5 we train five different AVM models to predict house prices based on transaction data for the city of Graz in Austria for the period 2015–2020. In particular, we train the following models: a (hedonic) linear regression model which serves as our parametric benchmark, a Random Forest model, a model with Multivariate Adaptive Regression Splines (MARS), a quantile regression model with LASSO penalties, and a simple Neural net model.Footnote3 We illustrate how the ranking of model performance varies depending on which metrics are used. Our final contribution is to suggest a short list of seven metrics (one from each of our seven classes) for general model evaluation. Our main findings are summarised in the conclusion in section 6.

2. AVMs – some general comments

2.1. AVM applications

AVMs in the real estate sector are algorithms (generally a combination of ML techniques) that are trained on large amounts of data in order to predict the current values of residential properties. They have become very popular in recent years, as they provide customers with a fast and simple way to compare real estate properties and monitor the market. AVM algorithms make use of many types of data, such as location, property age, or condition, and are able to generate a report within seconds. They are convenient for potential buyers of real estate as they minimise the need to personally inspect each property on the market.

But AVMs are not only used by buyers and sellers of real estate: They are also employed by banks to assist with mortgage lending, by insurance agencies to help with risk assessment, and by taxation offices to establish appropriate property tax levels. More and more AVM providers are entering the market. Even in a small country like Austria there are several AVM providers (some have entered the Austrian market after first establishing themselves in neighbouring countries). However, even though AVMs are increasingly used, it is hard to get information on how they are structured and on how they perform. For the user, an AVM is basically a black box.Footnote4

For big customers like banks or insurance agencies, the metrics in this paper could potentially be used to compare the performance of alternative AVM providers. To do this they could simply get a number of real estate price predictions from competing AVM providers and then construct some of the metrics we discuss in this paper. In particular, we discuss a short list of seven metrics that should provide a good AVM performance overview (see section 5.2).

2.2. Loss functions

We illustrate the importance of metrics in choosing between different AVM algorithms by training five different parametric, semi-parametric and non-parametric models. These are:

Model M1: Linear regression (hedonic) model as the parametric benchmark model

Model M2: Random Forest algorithm (non-parametric)

Model M3: Multivariate Adaptive Regression Spline Model (MARS) (semi-parametric)

Model M4: Quantile Regression with LASSO penalty (parametric)

Model M5: Neural Network algorithm (non-parametric)

We describe these algorithms in more detail in Appendix B.

Before we discuss which metrics to use for judging ML model performance, it is useful to consider the relationship between metrics and loss functions. Both loss functions and metrics are necessary to build good AVMs: loss functions to optimise algorithms while training the model, and metrics to judge their performance.Footnote5 When building a model, the decision of which loss function to use should be made together with the decision on the model type. However, when comparing the performance of multiple ML models, it is very likely that the models under consideration were trained using different loss functions. Thus, performance metrics are needed to compare the output of competing models. The metrics we describe in this paper are applicable in such situations.

Here, we decided to use quadratic loss (also called squared error loss or L2) to train all ML models. We do this to keep the structure of our paper as homogeneous as possible and to focus on the metrics rather than the loss functions.Footnote6

2.3. Cross validation

Model selection between parametric models is primarily centred around variations on the Akaike Information Criterion (AIC; Akaike, Citation1973), where to avoid overfitting the use of more parameters is penalised. On the other hand, model selection for ML models is generally done via cross-validation (CV; Yang, Citation2007). CV refers to various types of out-of-sample testing techniques (e.g. delete-1 CV, k-fold CV, etc.) that – to prevent overfitting – split the dataset (once or multiple times) and then use part of it to fit a particular model and the rest of the data to measure its performance. If the dataset has been split into k parts, the overall test error is estimated by taking the average test error across k trials (Goodfellow et al., Citation2016). Yang (Citation2007) shows that CV can become a consistent criterion for selecting the better procedure with probability approaching 1. One prerequisite for using CV as a consistent model selection criterion is that the dataset is sufficiently large (see Li (Citation1987) and Yang (Citation2007)).

We use CV for model specification (the procedure to find tuning parameters for each model type) as well as for model selection (the overall discrimination between the different parametric and semi-parametric and non-parametric models). As the goal of AVMs is to predict the prices of new unseen properties, we chose a growing window approach for the model-selection CV in which we always choose the latest data as test-set (as discussed in Cerqueira et al., Citation2019). See section 4.5 for more information on how we partition the data.

Once we decided on CV as model selection technique, we also need metrics to judge the performance of competing models. In this paper, we rationalise the relevant literature by collecting and classifying a wide range of metrics that can be chosen for this task and illustrate how the choice of metric can significantly influence which model is chosen via the CV process.

3. Performance metrics

3.1. What properties should a metric have?

Due to the inter-disciplinary nature of the literature, the terminology used varies across articles. In what follows, we try to use the most common terminology. Given that our focus is on the prediction of real estate prices, we will call our realised values pn and our predictions pˆn, where n=1,,N indexes the real estate units in the dataset. To ease interpretation and comparability we formulate (or re-formulate) all metrics so that lower absolute values indicate better model performance.

When comparing metrics, one property we will refer to is that of symmetry. We consider two versions of symmetry defined here as follows:

Symmetry-in-bias: A metric M is symmetric in bias if Mp,pˆ=Mpˆ,p.

In other words, a metric is symmetric in bias if swapping the actual and predicted prices changes the sign of the metric but not its absolute value. Thus, a prediction method is biased if there exists a systematic (positive or negative) difference between the actual observed prices and the predicted prices. Average bias metrics are designed to detect such bias; we present them in section 3.2.

Symmetry-in-dispersion: A metric M is symmetric in dispersion if Mp,pˆ=Mpˆ,p.

In other words, a metric is symmetric in dispersion if it is unaffected by swapping the actual and predicted prices. Metrics that violate symmetry-in-dispersion do not treat errors from the right tail in the same way as those from the left tail. We will illustrate this problem in more detail in section 3.5.

In what follows we consider seven classes of metrics that are (or could be) used to evaluate the performance of AVMs. They are: Average-bias metrics, Absolute-difference metrics, Squared-difference metrics, Absolute-ratio metrics, Squared-ratio metrics, Percentage-error metrics, and Quantile metrics. The relevance of the symmetry conditions will be discussed in the context of each class of metric.

Absolute metrics (sections 3.2 and 3.4) have the advantage over squared metrics (sections 3.3 and 3.5) that they are less sensitive to observations where there is a large discrepancy between the actual and predicted price. Which is better out of difference metrics (sections 3.2 and 3.3) and relative metrics (sections 3.4, 3.5, 3.6 and 3.7) depends on the context. Sometimes it is more important to measure prediction errors in monetary units while in other situations it is the percentage error that matters more.

3.2. Average bias metrics

Average bias metrics can be positive or negative. The closer the metric is to zero, the better is the performance of a method. This is in contrast to all the other classes of metrics considered later, which by design cannot take negative values.

Some authors in the literature have argued that certain users may actually prefer biased metrics. For example, Shiller and Weiss (Citation1999) note that when valuating properties for mortgage loans, a bank may prefer conservative predictions that have a downward bias. Similarly, for property tax assessments, Varian (Citation1974) suggests that a downward bias might also be desirable, so as to generate less complaints. More generally, buyers may prefer a downward bias, and sellers an upward bias.

However, in our experience, customers want AVM output to be as accurate and bias-free as possible. Risk adjustments to the AVM predictions can subsequently be made as required. For example, banks generally reduce the valuation from AVMs before deciding on mortgage loans. Banks and other financial institutions prefer to make these risk adjustments themselves rather than relying on risk-adjustments that emerge in an opaque manner directly from the AVM.Footnote7 The same principle applies to property tax assessments. Even individual buyers and sellers can make their own adjustments to meet their needs. One exception to this principle is that firms valuing their collateral may prefer an upward biased AVM to increase their share price or to get easier access to bank loans (see Agarwal et al., Citation2020).

The two main average bias metrics are MBE and MDBE.

Mean Bias Error (MBE):

MBE=1Nn=1Npnpˆn.

Median Bias Error (MDBE):

MDBE=medpnpˆn,

where med refers to the median of the prediction errors.

MBE is used by Enström (Citation2005) to assess whether appraisal overvaluation contributed to the severe property crisis in Sweden in the early 1990s. Schulz et al. (Citation2014) use both MBE and MDBE to evaluate model performance.

Mean and median prediction error ratio metrics like MPE, MPE’, and MDPE can also be interpreted as bias metrics. These metrics are presented in . MBE and MDBE satisfy symmetry-in-bias, while MPE, MPE’, and MDPE do not. Replacing the error ratio minus 1 by the log error ratio in MPE and MDPE yields the metrics LMPE and LMDPE (also shown in ). Both LMPE and LMDPE satisfy symmetry-in-bias.

Table 1. Average bias metrics

The symmetry-in-bias property is important when measuring average bias, since the ordering of p and pˆ is arbitrary in a metric formula. For example, while it is more standard to write the error ratio as p/pˆ, it could equally well be formulated as pˆ/p. If symmetry-in-bias is violated, then an average bias metric will be affected by this arbitrary ordering, potentially leading to an altered ranking of ML methods.

3.3. Absolute-difference metrics

The remaining metrics all focus on measuring the average dispersion error, without attention to bias. The smaller the average error, the better is the method. We turn first to absolute-difference metrics. These metrics measure the average (mean or median) absolute error.Footnote8 Absolute-difference metrics limit the impact of individual outliers on model performance (in comparison to squared-difference metrics discussed below) and are therefore particularly useful in situations where data entry errors or other data quality issues are a problem. They are good model selection criteria when (repeated) average performance is important.

Mean Absolute Error (MAE):

MAE measures the average of the sum of absolute differences between observation values and predicted values and corresponds to the expected loss for the L1 loss function.

(1) MAE=1Nn=1Npnpˆn.(1)

Median Absolute Error (MDAE):

MDAE=medpnpˆn.

Both MAE and MDAE satisfy symmetry-in-dispersion. MAE is used in an AVM context by Zurada et al. (Citation2011), McCluskey et al. (Citation2013), Masias et al. (Citation2016), and Yacim and Boshoff (Citation2018). MAE is also used by Smith et al. (Citation2007) to measure the accuracy of temperature predictions. Diaz-Robles et al. (Citation2008) use MAE to predict particulate matter levels in urban areas. See also for the formulas absolute difference metrics.

Table 2. Absolute-difference metrics

3.4. Squared-difference metrics

An alternative to averaging absolute errors (as in section 3.3) is to average squared errors. Squared-difference metrics are more sensitive to outliers than absolute-difference metrics. They are particularly useful for situations where large prediction errors need to be minimised.Footnote9

The mean squared error (MSE) of an estimator measures the average squared difference between the estimated and true values. MSE is a risk function, corresponding to the expected value of the squared error loss L2.

Mean Squared Error (MSE):

(2) MSE=1N n=1Npnpˆn2(2)

The Root Mean Squared Error (RMSE) is a monotonic transformation of the MSE, which is commonly used in the literature. We prefer RMSE over MSE since it generates smaller values that are more easily compared across methods and hence easier to interpret for the user.

Root Mean Squared Error (RMSE):

(3) RMSE=n=1Npnpˆn2N,(3)

MSE and RMSE are symmetric in dispersion.

provides an overview of the squared difference metrics discussed here. There are also a number of metrics that build on squared error loss, but generally loose the symmetry-in-dispersion property. We list some of these alternative squared-difference metrics in . All of these metrics (R2, CC, NRMSE, SNR, and SDE) violate symmetry-in-dispersion.

Table 3. Squared-difference metrics

Table 4. Absolute ratio metrics

Table 5. Squared-ratio metrics

One of them is the widely used metric R2:

1R2=n=1Npnpˆn2n=1Npnpˉ2,

where pˉ is the arithmetic mean of the observed prices. We use 1 − R2 as our performance metric so that smaller values are better, which makes this measure comparable with the other metrics considered here.

In the literature, MSE and R2 are used to evaluate AVM performance by Kok et al. (Citation2017), Masias et al. (Citation2016), Bogin and Shui (Citation2018), and Yacim and Boshoff (Citation2018) all use RMSE and R2 Peterson and Flanagan (Citation2009), Zurada et al. (Citation2011), and McCluskey et al. (Citation2013) all use RMSE.

Squared-difference metrics have also been used in other contexts. Abdul-Wahab and Al-Alawi (Citation2002) use R2 to predict ozone levels. Bajari et al. (Citation2015) use RMSE to compare methods of predicting grocery store sales. Wu et al. (Citation2004) use RMSE to compare methods of predicting travel times. In a survey paper on forecasting wind power generation, Foley et al. (Citation2012) discuss R2, NRMSE, and CC as possible metrics. Spüler et al. (Citation2015) use NRMSE, CC, and SNR to evaluate methods of decoding neural signals in the brain.

3.5. Absolute-ratio metrics

In sections 3.3 and 3.4, prediction errors were measured as differences between actual and predicted values.Footnote10 However, prediction errors can also be measured as ratios. In many situations, ratio-based measures are more relevant. For example, a 10,000 USD error on a house that sold for 100,000 USD will often be viewed as worse than a 10,000 USD error on a house that sold for one million dollars. As was the case with difference errors, ratio errors can also be measured in either absolute value or squared form; again, absolute-ratio metrics are less sensitive to outliers. Focusing first on absolute error ratios, three popular metrics are defined below:

Mean Absolute Prediction Error (MAPE):

MAPE=1Nn=1N|ppˆn1|

Median Absolute Prediction Error (MDAPE):

MDAPE=med |ppˆn1|

Coefficient of Dispersion (COD):

COD=1Nn=1Npnpˆn/medppˆn1.

MAPE, MDAPE and COD are not symmetric in dispersion. Hence by reversing p and pˆ we obtain different metrics which we list under MAPE’, MDAPE’, and COD’ in .

To illustrate the problems that can arise when a metric violates symmetry-in-dispersion, it is informative to consider the case of MAPE. When pˆp, p/pˆ is unbounded. It can take any value from 1 to infinity. Conversely, when pˆp, p/pˆ is bounded to lie between 0 and 1. Hence errors in the right tail have only a limited impact on MAPE, while errors in the left tail can potentially have a huge impact. This asymmetry in the treatment of errors in the left and right tail can distort MAPE in unexpected ways, potentially producing a misleading ranking of ML methods.

Nevertheless, MAPE is one of the most widely used metrics in the AVM literature. For example, it is used by D’Amato (Citation2007), Peterson and Flanagan (Citation2009), Zurada et al. (Citation2011), Schulz et al. (Citation2014), Ceh et al. (Citation2018), and McCluskey et al. (Citation2013) use both MAPE and COD. COD is used by Moore (Citation2006) and Yacim and Boshoff (Citation2018). Moore states that COD is widely used to measure quality in the tax assessment field (see also Stewart, Citation1977). An alternative definition of COD replaces the median with the arithmetic mean (see Spüler et al., Citation2015). This example illustrates the importance of providing a precise formula to avoid confusion (particularly given the interdisciplinary nature of the literature).

Makridakis (Citation1993) and Hyndman and Koehler (Citation2006) show how MAPE and MDAPE can be modified to make them symmetric:

Symmetric Mean Absolute Percentage Error (sMAPE):

sMAPE=1Nn=1Npnpˆnpn+pˆn .

Symmetric Median Absolute Percentage Error (sMDAPE):

sMDAPE=medpnpˆnpn+pˆn.

These symmetric metrics address the dilemma over which of pn and pˆn should go in the denominator by adding them together and putting both in the denominator.

We propose here three other symmetric variants on MAPE that are new to the literature: First, symmetry can be imposed by replacing pnpˆn−1 by ln(pn/pˆn).Footnote11 This provides us with the following metric:

Log Mean Absolute Prediction Error (LMAPE):

LMAPE=1Nn=1Nlnpnpˆn.

A second way of imposing symmetry is by replacing pnpˆn−1 by max(pn, pˆn)/min(pn, pˆn) −1).

Max-Min Mean Absolute Prediction Error (mmMAPE):

(4) mmMAPE=1Nn=1Nmaxpn,pˆnminpn,pˆn1.(4)

A third way of imposing symmetry is by replacing pnpˆn1 by pnpˆn+pˆnpn2. The resulting metric corresponds to the first of three metrics that Diewert (Citation2002, Citation2009) proposes for measuring the dissimilarity of price vectors across time periods or countries:

Diewert Metric 1 (DM1):

D1=1Nn=1Npˆnpn+pnpˆn2 .

All three of Diewert’s metrics are well suited to measure the dissimilarity between actual and predicted values in AVMs – especially since they all satisfy the symmetry-in-dispersion criterion. As far as we are aware, these three metrics are new to the AVM literature. The second and third metrics – DM2 and DM3 – belong to the Squared-ratio class and are discussed below in section 3.6.

3.6. Squared-ratio metrics

The most widely used squared-ratio metric is MSPE, defined below:

Mean Squared Prediction Error (MSPE):

MSPE=1Nn=1Npnpˆn12.

The MSPE is used by Schulz et al. (Citation2014). This metric is not symmetric, which implies that if p and pˆ are reversed, we obtain a different metric (see ).

Squared-ratio metrics that violate symmetry-in-dispersion suffer from an even more severe version of the criticism of MAPE discussed in section 3.5. Prediction errors in the left and right tail of the error distribution will be weighted in a highly asymmetric and potentially misleading way. Squaring the errors acts to amplify this effect.

We consider three symmetric variants on MSPE. The methods for imposing symmetry-in-dispersion are essentially the same as those applied to MAPE above.

First, replacing pnpˆn1with ln(pn/pˆn) we obtain the following:

Log Mean Squared Prediction Error (LMSPE):

(5) LMSPE=1Nn=1Npnpˆn2 ,(5)

Interestingly, this metric turns out to be identical to Diewert’s (Citation2002, Citation2009) third metric.

Second, replacing pnpˆn1by max (pn, pˆn)/min (pn, pˆn) −1, turns MSPE into:

Max-Min Mean Squared Prediction Error (mmMSPE):

(6) mmMSPE=1Nn=1Nmaxpn,pˆnminpn,pˆn 12(6)

Third, replacing pnpˆn1 by (pnpˆn−1) + (pˆnpn−1) transforms MSPE into:

Diewert Metric 2 (DM2):

DM2=1Nn=1Npˆnpn12+pnpˆn12 .

3.7. Percentage error ranges

The percentage error range (PER) counts the percentage of prediction error ratios that lie outside a specified limit.Footnote12 In this sense, PER is related to the concept of Value-at-Risk from the finance literature (see for example, Dowd, Citation2005). We consider a few versions of this type of metric.

Percentage Error Range (PER):

PERx=100pnpˆn1>x.

To understand PER, we consider an example in which we set x = 10 and assume the answer is PER(10) = 40. This tells us that the error rate is above 10 percent in 40 percent of the valuations.

According to Crosby (Citation2000), metrics of this type are often used as benchmarks for expert witnesses in court cases involving the valuation of real estate assets. More specifically, an expert’s estimate is expected to not deviate by more than say 10 percent from the actual current market value.

PER is not symmetric in dispersion, thus by reversing p with pˆ we obtain PER’ (see ). However, by applying the log- or max-min transformations as before, we can obtain two symmetric versions of PER:

Table 6. Percentage-ratio metrics

Log Percentage Error Range (LPER):

LPERx=100lnpnpˆn>x.

Max-Min Percentage Error Range (mmPER):

mmPERx=100maxpn,pˆnminpn,pˆn1>x.

3.8. Quantile metrics

Some additional metrics that do not fit into any of the classes discussed above are included above in . The extent to which model performance is robust with respect to extreme values is an important consideration in many ML implementations. Median based metrics (such as MDPE) are more robust measures of central tendency in the error distribution than means (see for example, Wilcox & Keselman, Citation2003). Similarly, the interquartile and 90–10 quantile ranges are more robust measures of dispersion than variance-based measures such as MSPE, RMSE, or mean absolute deviation measures such as MAPE. Hence quantile-based metrics for measuring the dispersion of the error distribution are useful additional diagnostic tools. Such metrics can be defined on the prediction errors measured as ratios or in levels, as shown in . All four metrics in are symmetric in dispersion.

Table 7. Quantile metrics

4. The prediction framework

4.1. The dataset

To illustrate our analysis, we estimate an AVM for apartments in Austria’s second largest city, Graz. Our data consists of all residential transaction data for the city of Graz for the time period January 2015 to April 2020. We were provided with this dataset by ZTdatenforum, a firm that transcribes every contract listed in the official Austrian deeds book (Grundbuch) into a transaction price dataset. The following variables are available for each transaction: actual transaction price, time of sale, internal space in square metres, balcony (yes/no), parking (yes/no), outside space (garden or terrace) (yes/no), private sale or purchase directly from builder, zoning classification determining the maximum allowed building density, age category (1–5), longitude and latitude. In total there are 18 957 transactions.

4.2. Cleaning

Before we fit the data to the different machine learning algorithms, we have to apply several pre-processing steps to improve the overall performance of the price valuations and to ensure comparable results among the different methods. We therefore first remove all non-market transactions, marked either by transactions within family (relationship), or due to insolvency. These transactions accounts for about 3.5% of all observations. Next, we consider only residential apartments and hence, remove all houses, offices, and attics from our dataset. In 27% of our data the information of geographical coordinates or living area is missing. As the purpose of this paper is to illustrate methods rather than establish a ‘complete’ AVM, we decided to exclude these transactions from our illustration here.Footnote13

4.3. Outlier detection

A total of 17 apartments are excluded from the dataset by restricting living areas to between 20 and 150 square metres (between 215 and 1615 square feet) and excluding those outside the price range of 10 000 and 1 500 000 EUR.

We then use an Isolation Forest – an automated anomaly detection algorithm based on machine learning (Liu et al., Citation2008) – to detect outliers. The Isolation Forest uses a random sub-sample of the training data, and splits each variable randomly until the pre-defined number of rounds is reached. After 1 000 rounds we observe the average number of splits needed to isolate each individual observation, sort the dataset according to this anomaly score, and delete the worst performers (4% of observations). illustrates the results of the isolation forest for longitude and latitude as well as price and square metre outliers. The graphs on the left illustrate the full dataset (the darker the colour the higher the anomaly score), while the figures on the right of show the results after outlier removal.

Figure 1. Selected variables before Outlier Detection (left) and after Outlier Detection (right) with the Isolation Forest Algorithm

Figure 1. Selected variables before Outlier Detection (left) and after Outlier Detection (right) with the Isolation Forest Algorithm

shows a summary of the data before and after the cleaning.

Table 8. Different stages of the data cleaning process

4.4. Variable creation and transformation

ML algorithms perform better with more observations and more input variables per observation. We therefore construct a variety of variables from the existing dataset variables as well as additional data sources. For example, for each apartment we calculate the distance to the city centre and to the nearest school, and we establish yearly, quarterly, and monthly time variables from the dates of sale. To better describe neighbourhood effects, we establish new variables based on location clustering (we use a k-Means unsupervised learning algorithm for this step). Additionally, we include data on the noise level of nearby streets. For a summary of the individual characteristics, see in Appendix C.

We transform price into log price to make the target variable more normal. This is a standard technique to improve the overall performance of house-price prediction models. To re-transform the estimated log values at the end, we apply the smearing adjustment factor introduced by Duan (Citation1983). This adjusts the estimated price pˆ such that:

(7) Epˆ=φˆexpX βˆ=φˆexplnpˆ(7)

with φˆ=1nNn=1exp(εi)

where X is a vector of property i specific characteristics, βˆ is a vector containing the corresponding characteristic shadow prices, and lnpˆ is the predicted log price obtained from the ML model. N is the sample size, i the difference between observed and predicted values in log form (i.e. lnpiln(pˆi)), and φˆ the adjustment factor.Footnote14 Footnote15

Furthermore, we centre and scale all numeric variables to lie between 0 and 1. Scaling inputs helps to avoid situations where one or several features dominate others in magnitude and as a result the model hardly picks up the contribution of the smaller scale variables, even if they are strong. Some ML algorithms are more sensitive to this than others.Footnote16

4.5. Train-test split

To evaluate the different metrics, we use a growing window 5-fold cross validation as depicted in . Since, the aim of an AVM is to perform well on new unseen data, we use a growing window approach to judge each algorithm on future data (see e.g. Cerqueira et al., Citation2019). For each fold, the training data are used to find the best set of parameters (using cross validation) and then the metric is calculated by using the test set. The final error is then computed by the average over all 5 prediction sets according to each metric.

Figure 2. 1-step ahead train/test split

Figure 2. 1-step ahead train/test split

4.6. Model tuning and selection

We train the following prediction methods: Linear Regression (hedonic) model (M1), Random Forest model (M2), Multivariate Adaptive Regression Splines (MARS) model (M3), a model that uses Quantile Regression with LASSO penalties (M4), and a Neural Net model (M5).

We chose these methods because they are widely applied in the AVM literature and of similar predictive power. A short description of each of these methods can be found in Appendix A. We use ‘R’ (R Core Team, Citation2012) to perform all computations.Footnote17

Most of the models we consider need some degree of model tuning to find the optimal hyperparameters. This involves making comparisons between different model versions. We use grid-searches on hyper-parameters and 10-fold cross-validation on the training set to select between model variations of one model family.Footnote18

We use the RMSE metric to measure performance in the tuning stage. This raises an important point: Performance metrics are needed at two stages in the modelling process. First, they are needed to tune the individual ML methods. Second, they are used to compare performance across different ML methods. While we focus here on the use of metrics in the second stage, similar issues arise in the first stage. Here it is practical for us to focus on a single metric during the first stage, so as to obtain unambiguous results when tuning the model. One attraction of RMSE in this regard is that it is a standard tuning metric which can often be taken ‘off-the-shelf’ in ML estimation packages. If it is applied to the prices in log form then it corresponds to our LRMSE metric (which satisfies symmetry).Footnote19 The result of this exercise is the model specification that then gets evaluated in the final stage via the test set (hold-out sample).

We use the metrics described above to compare the performance of five ML prediction methods. Performance across methods is then evaluated using the holdout sample. We use all 48 metrics discussed above to make this evaluation, and compare how sensitive the results are to the choice of metric.

5. Metric results and implications

5.1. The sensitivity of rankings to the choice of performance metric

Each apartment i in the test set is characterised by a price pi and a set of characteristics Xi=xi,1,xi,2,,xi,d, where xi,j denotes the jth characteristic of the ith apartment. Each of the trained models (M1 to M5) is provided with the set Xiand generates predictions for the logarithm of the price lnpi for each apartment.i.

illustrates the distribution of the log errors for each model.Footnote20 The black solid line corresponds to the normally distributed error density derived from the mean and standard deviation of the observed errors. The skewness scores indicate how skewed a distribution is (a skewness score of zero represents a symmetric distribution). All our models have moderate positive skewness scores, implying they have larger overpredictions than underpredictions (in log-difference form). The Random Forest model (M2) is least affected by skewness.

Figure 3. Error distribution of model M1 – M5

Figure 3. Error distribution of model M1 – M5

All metrics discussed in this paper are calculated using the set of true prices (pi) and the set of predicted prices (pˆi); lists the output for all 48 metrics.

In addition to these average cross-validation results, we also present the best model for each individual fold in . illustrates that the overall performance is stable across the different sub-datasets. M2 (Random Forest) performs best for 30 metrics, M3 (Multivariate Adaptive Spline Model) is best for 10 metrics, M4 (Quantile Regression with LASSO Penalty) is best for 5 metrics, M5 (Neural Net) is best for 2 metrics, and M1 (Linear Regression) is best for 1 metric.

An important additional consideration to the overall number of ‘wins’ by individual models is the pattern that emerges. While M2 (Random Forest) dominates overall, it does not do so for metrics in the Mean Bias class or the Squared-Ratio class. For these classes M3 (Multivariate Adaptive Spline Model) outperforms M2 (Random Forest).

5.2. What set of metrics do we recommend?

It is not always practicable to calculate the performance of alternative models for nearly 50 metrics. Hence there is a need to prioritise and decide which metrics are most important.

lists some of the most commonly used metrics in the AVM literature: Mean Prediction Error (MPE), Mean Absolute Prediction Error (MAPE), Coefficient of Dispersion (COD), Root Mean Squared Error (RMSE), Mean Absolute Error (MAE), and R2. Using these metrics, the model of choice is M2 (Random Forest) for all these metrics except MPE (where model M4 performs best).

Table 9. Model performance rankings based on metrics from the AVM literature

In we present seven metrics (one for each class) that we believe provide a better basis for evaluating AVM performance. It is important to select a mix of metrics that capture different aspects of model performance.

Table 10. Predictive performance of methods M1-M5 (Short List)

LMDPE is selected as a measure of central tendency (bias) in preference to MPE in since it is symmetric in bias and more robust to outliers. The greater robustness follows from LMDPE being median based.

MAE, mmMAPE, LRMSE, RMSE, mmPER and IQRAT all measure dispersion. Both MAE and RMSE satisfy symmetry-in-dispersion and we keep them to represent the absolute-difference and the squared-difference metrics. Both of these metrics are also widely available in statistical packages. Note that as an absolute-difference metric MAE is more robust to outliers than RMSE. Thus, depending on the dataset and/or task at hand, one would give one or the other more weight.

The two absolute-ratio metrics in , MAPE and COD, do not satisfy symmetry-in-dispersion. We replace them with mmMAPE, which is a symmetrified version of a number of different widely used metrics, of which MAPE is just one example.Footnote21

The squared-ratio metrics, range metrics, and quantile metrics were not represented in . We choose LRMSE as the representative for the squared ratio class. It satisfies symmetry in-dispersion and is also easily calculated with most statistical packages (by simply applying RMSE to the log prices).

Our representative for the range metrics is mmPER(10), while IQRat is chosen as the representative for the quantile metrics. The results for these metrics are shown in . The metrics LMDPE, mmPER and IQRAT are new to the literature.

In , M2 (Random Forest) performs best according to MAE, mmPER(10) and IQRAT, while M3 (Multivariate Adaptive Regression Spline) performs best according to MDPE, and LRMSE. For LMDPE, the difference in the results between M3, M4 and M5 is very small. All three medians are closely centred on the target value of zero. Similarly, the difference in performance for LRMSE between M3 and M2 is minimal. Hence based on this short list, M2 performs best in terms of average dispersion, although it has a slight tendency to overpredict the median error ratio.

6. Conclusion

The choice of performance metric to compare alternative ML models is a potential source of confusion in the AVM literature. A number of metrics are used, but there has been little attempt to undertake a systematic analysis of their properties. We collect 48 different metrics and structure them according to their mathematical formulation into seven classes. These are: the class of bias metrics, two ways of constructing difference metrics (absolute and squared differences), two ways of constructing ratio metrics (absolute and squared ratios), metrics based on percentage-ranges, and quantile-based metrics. This systematic review and classification of existing metrics is one of the contributions in this paper.

Users will find metrics from one or the other class more appropriate depending on how they want to treat errors. However, not all metrics of a class are equally well suited as performance metrics for AVMs. For example, some very popular metrics do not treat prediction errors in a symmetric manner. We differentiate between two types of symmetry that are important in this context: symmetry-in-bias (important for average-bias metrics) and symmetry-in-dispersion (important for the other classes). Failure to satisfy these symmetry criteria implies some arbitrariness in the metric formula. For example, it can lead to asymmetric treatment of the left and right tails of the error distribution. Such asymmetries can distort results in unexpected ways, potentially producing a misleading ranking of ML methods.

Another contribution of this paper is the transformation of some commonly used metrics that violate the symmetry conditions. In this way, we introduce a number of new metrics into the literature. These new metrics still measure the same type of error as before, but do this now in a symmetric way.

Our empirical applications illustrate the need to think carefully about which set of metrics should be used to choose between models. We evaluate five ML models with 48 different metrics and find that each model performs best at least once (depending on the chosen metric). Thus, arbitrarily choosing a metric can lead to an inappropriate model being chosen. We provide some guidance in this matter by presenting a shortlist of seven metrics – one for each class. Each of these metrics addresses a different aspect of the performance evaluation problem; taken together, this shortlist will provide a good overall evaluation of alternative AVM models. Similarly, if a user is mainly interested in one class of errors, the shortlist provides an ideal candidate.

The following three ingredients are necessary to build effective ML prediction models according to Kuhn (Citation2016):

(1) intuition and deep knowledge of the problem context,

(2) relevant data, and

(3) a versatile computational toolbox of algorithms.

We recommend adding another ingredient to this list:

(4) the appropriate choice of performance metrics for model selection via cross-validation.

Acknowledgments

We acknowledge financial support for this project from the Austrian Research Promotion Agency (FFG), grant #10991131.

We also thank ZTdatenforum (www.zt.co.at) for providing us with the dataset.

Disclosure statement

No potential conflict of interest was reported by the authors.

Additional information

Funding

This work was supported by the Österreichische Forschungsförderungsgesellschaft FFG [grant #10991131].

Notes on contributors

Miriam Steurer

Miriam Steurer is an economist at the Institute of Economics at University of Graz, Austria. She obtained her PhD from University of New South Wales in Australia in 2008. Her current research focuses mainly on the application of machine learning methods to the real estate market. As part of this research agenda, she is developing Automated Valuation Models for the Austrian real estate market in partnership with the firm ZTdatenforum. This project is funded by the Austrian Research Promotion Agency (FFG). Dr Steurer is included in the expert roster of the IMF Statistical Department and was part of the Technical Advisory Group to the 2017 International Comparison Program of the World Bank.

Robert J. Hill

Robert J. Hill is Professor of Macroeconomics at University of Graz, Austria. Before that he was Professor of Economics at University of New South Wales in Australia. He obtained his PhD in Economics in 1995 from University of British Columbia in Canada. He is a Fellow of the Academy of the Social Sciences of Australia, the Society for Economic Measurement, and a member of the Conference on Research in Income and Wealth at the NBER. He is currently serving on World Bank and UN task forces. He has previously served on task forces for the UK Statistical Authority and Eurostat. From 2008 to 2014 he was Managing Editor of the Review of Income and Wealth. He has published widely in journals such as American Economic Review, International Economic Review, Review of Economics and Statistics, Journal of Econometrics, Journal of Business and Economic Statistics, Economic Inquiry, Regional Science and Urban Economics, Journal of Housing Economics, and Journal of Real Estate Finance and Economics.

Norbert Pfeifer

Norbert Pfeifer is a PhD student in the Institute of Economics at University of Graz; Austria (supervisor: Prof. Hill). In 2019 he presented papers at both the American Real Estate and Urban Economics Association (AREUEA) conference at Bocconi University in Milan, and at the European Real Estate Society (ERES conference at Cergy-Pontoise in France. His main research area lies in the application of computer vision and natural language processing applied to the field of housing. He is also engaged in the construction of an Automated Valuation Model for the Austrian real estate sector.

Notes

1. This point has been stressed both in the academic literature (see Varian, Citation2014) as well as in practical applications (e.g. the winning entries of Kaggle competitions (www.kaggle.com/competitions) almost invariably use ML methods).

2. Sometimes the term ‘Mass Appraisal’ is used instead of AVM in the literature.

3. A discussion on these 5 algorithms can be found in Appendix B.

4. For example, none of the AVM providers in Austria publish information on their algorithms, their objective functions, or performance metrics.

5. Some of the metrics themselves can be derived from loss functions.

6. Although in two of the models (models M3 and M4) we added penalisation terms to the loss function to allow for variable selection and to minimise over-fitting of the ML model.

7. We were told that this is the approach followed by banks in Austria.

8. Underpinning this class is the absolute loss function: L1=|ppˆ|.

9. Underpinning this class is the squared loss function: L2=ppˆ2

10. Either in absolute or squared terms.

11. It is worth noting that pnpˆn−1 is a first order Taylor series approximation of lnpn/pˆn

12. PER measures the expected loss for a 10 loss function, where L = 1 if the ratio error is greater than x, and L = 0 otherwise.

13. Alternatively, we could have imputed the missing characteristics based on their statistical attributes (such as mean, or median values) or by using more sophisticated methods (e.g. missing forest or k-nearest-neighbours algorithms).

14. This Jensen-type correction is needed because E[pˆ]=E[exp (Xβˆ+ϵ)]exp(Xβˆ).

15. Given that we are training the random forest model on actual prices rather than log prices (no adjustment is needed) and the applied neural network with one hidden layer relies on a linear relationship between the predicted price and the observed inputs – as for the remaining methods – Equationequation (7) can be used for the re-transformation adjustment without any further assumption.

16. In particular neural network algorithms do not have the property of scale invariance (see e.g. Hastie et al., Citation2009).

17. There are various packages in ‘R’ that can be used to train ML algorithms. For many ML related processes, the ‘caret package’ (Kuhn, Citation2008) is a starting point as it provides many different ML techniques in one comprehensive framework.

18. Note, we completely retrain all models after each new train-test split.

19. When LRMSE is used to tune the models in the first stage, then to ensure internal consistency LRMSE should also be one of the metrics used to compare performance across ML methods in the second stage.

20. We take the log errors from fold 5 to construct the histograms in .

21. Note that LMAPE or DM1 could also play this role.

22. For a survey of this literature see Hill (Citation2013).

23. See Diewert (Citation2003) and Malpezzi (Citation2008) for a discussion of some of the advantages of the semilog functional form in a hedonic context.

24. We follow Hastie et al. (Citation2009) and first create linear basis functions with a ‘knot’ at each observed input value xi,j, such that for each xi,j we have a function-pair of the form:max0,Xjxi,j and max0,xi,jXj (Hastie et al., Citation2009). These piece-wise linear basis functions can then be transformed by multiplying them together (which can form non-linear functions).

25. By adding the L1 penalty term to the Error term, the LASSO exploits the bias-variance trade-off to produce models that increase the bias in the model in order to greatly reduce the model variance and thereby combat the problem of collinearity (Kuhn, Citation2016).

26. Note that for each quantile τ the solution to the minimisation problem yields a separate set of regression coefficients.

27. Sherwood (Citation2017) provides an R-package called ‘rqPen’ which implements (12).

28. The error rate is defined as the difference between the Neural Net prediction and the observed transaction price.

References

Appendix A.

Results

Table A1. Predictive performance of methods M1-M5 (Full-List)

Table A2. Cross validation results

Appendix B.

Description of the applied models

Model M1: Linear Regression

The Linear Regression model (hedonic model) serves as a benchmark for the ML models. Hedonic models have originally been used in economics to model the prices for products subject to rapid technological change, such as cars and computers (see for example, Grilliches, Citation1961). They have since become the standard model for house price regressions, especially when it comes to house price indices.Footnote22 The hedonic model regresses the price of a product on a vector of characteristics, whose prices are not independently observed, thereby generating shadow prices for these characteristics. If the logarithm of the price is used on the left hand side, the interpretation changes slightly: instead of shadow prices on characteristics, the estimated parameters then estimate the percentage influence of these characteristics.Footnote23 Here we regress the logarithm of price on a linear function of explanatory characteristics. All categorical variables are included as dummy variables.

Thus, the model parameters (the βs) are chosen to minimise the Sum-of-Squared Errors (SSE):

minβiN(yiβ0+βxi)2,

where β0 indicates the intersect term.

Model M2: Random Forest

The random forest technique was first proposed by Breiman (Citation2001) and has since become one of the most popular ML methods. For house price predictions they have first been used by Antipov and Pokryshevskaya (Citation2012).

Random Forests are tree-based non-parametric ensemble methods with uncorrelated decision trees as base learners. Each tree is a simple model that is built independently using a random sample of the available variables. Averaging over many independently built trees reduces variance, increases robustness, and makes the method less prone to over-fitting. Different techniques exist to construct base learners (the individual trees). The most common one is the ‘Classification And Regression Tree’ (CART) method, also known as the recursive partitioning procedure which was proposed by Breiman et al. (Citation1984).

Building a CART tree begins by splitting the dataset S into two groups (S1 and S2) so that the overall sum of squared errors (SSE) are minimised. To find the predictor and split value that minimises the SSE, it tries out every distinct value (split point s) of every predictor (Kuhn, Citation2016). Thus, for each variable j and each split point s, we minimise the following:

minj,siεS1(yiyˉ1)2+iεS2(yiyˉ2)2,

where (yˉ1) and (yˉ2) denote the averages of the target values of S1 and S2 respectively. This process is then repeated within each subgroup, continuously splitting the data into smaller subsets.

A random forest builds an ensemble out of many such trees. Correlation between predictors is reduced by providing the algorithm with a randomly chosen number of predictors at each split (rather than the full set of available predictors). The number of predictors presented to the algorithm at each step is generally referred to as ‘mtry’ and is the main tuning parameter of the random forest model. The other tuning parameter is the number of individual trees that are grown and averaged. We use grid searches over these tuning parameters and CV using the LRMSE metric to find the best performing version of the random forest model at this stage.

Random Forests are robust to outliers and a good method when data are noisy. A Random Forest model can consist of mixed variables (numerical and categorical), and/or contain missing values. These features – plus the fact that they are relatively easy to implement – have made Random Forest models very popular.

Model M3: Multivariate Adaptive Regression Splines (MARS)

An example of a non-parametric extension of linear regression is the multivariate adaptive regression spline (MARS), which was introduced by Friedman (Citation1991). The basic idea is as follows: ‘A piecewise polynomial function fx is obtained by dividing the domain of X into continuous intervals and representing f by a separate polynomial in each interval.’ (Hastie et al., Citation2009). Continuity constraints are introduced to make the resulting polynomial function fx continuous at the threshold points (knots). The ‘division’ of the domain X is done via hinge functions – linear basis functions that identify the threshold points (knots), where a linear regression model is shifted into a different regression line. The first step in the MARS algorithm is thus the formation of these hinge functions.Footnote24

Following the terminology in Hastie et al. (Citation2009), we can write the regression problem as follows:

fX=β0+m=1MβmhmX,

where each hmX represents a hinge function or a product of hinge functions.

The model building process of MARS is then like a stepwise linear regression using the basis functions – and their transformations – as inputs (Hastie et al., Citation2009).

The model coming out of this regression will be over-fitted and therefore needs to be trimmed back. This can be done via a stepwise term deletion procedure. One by one the terms, that are least helpful in reducing the overall error, are removed until only the intercept term remains. Each of these deletions leaves us with a possible model. We again use CV with LRMSE as the metric to select the one that fits our data best.

Regression splines provide a highly versatile regression technique that is relatively easy to implement and has many benefits: it automatically performs variable selection, variable transformation, and interaction detection. It also generally produces lower errors than linear regression techniques. Like linear regression, MARS is relatively easy to interpret, but is – because of the non-parametric parts – more flexible. Due to these benefits, applications of MARS are diverse and range from forecasting grid power output (Li et al., Citation2016) to droughts in Australia (Deo et al., Citation2017).

Model M4: Quantile Regression with LASSO Penalty

First introduced by Koenker and Bassett (Citation1978), Quantile Regression (QR) explicitly addresses a weakness of standard regression techniques: the focus on predicting in the vicinity of the mean of the distribution, and hence often poor performance away from the mean (Zou, Citation2006). LASSO penalties were introduced by Tibshirani (Citation1996) as a method to perform parameter shrinkage and parameter selection in linear regression models, but can also be applied to other statistical models. LASSO stands for ‘Least Absolute Shrinkage and Selection Operator’ and refers to a penalty in L1 norm of the coefficient vector. It is particularly well suited to problems with sparse data (Hastie et al. (Citation2009)).Footnote25

The combination of Quantile Regression with L1 norm shrinkage was first applied by Koenker (Citation2004). Wu and Liu (Citation2009) show that the inclusion of a LASSO penalty parameter can improve interpretability without losing accuracy in fit. He et al. (Citation2016) apply a Quantile regression model with LASSO penalties to identify genetic features that influence quantitative traits. We are not aware of any applications in the housing market of this method thus far.

Unlike least-squares regression, where the coefficients are estimated by solving the least squares minimisation problem, the coefficients in a linear quantile regression are chosen by minimising the sum of asymmetrically weighted absolute errors:

minβτi=1NρτyixiTβτ,

where τ refers to the individual quantile being modelled, and the weights ρτu are given by: ρτu=τuifu>0,and1τu otherwise.Footnote26

After adding the LASSO penalty term, (11) becomes:

minβτi=1NρτyixiTβτ+αj=1dβτ,j.

Tuning of the model is done by choosing the number of quantiles τ and the regularisation parameter α, which controls the strength of the shrinkage process (and thus also variable selection). Too much shrinking leaves a sub-optimal model, while too little shrinking tends to lead to poor interpretability (and over-fitting). Again, for model selection in the tuning phase, we use the LRMSE metric and cross-validation to choose the best-performing regularisation parameter.Footnote27

Model M5: Neural Nets

Neural Net models have a wide variety of applications, most notably in speech recognition and machine translation, computer vision (object and activity recognition), and robotics (e.g. self-driving cars). They are particularly useful when automated feature selection is needed and when the dataset is large. Our dataset, consisting of just under 6000 transactions and a limited number of variables, is rather small for a Neural Net application. Neural net applications to the estimation of house prices include Selim (Citation2009), Peterson and Flanagan (Citation2009), Zurada et al. (Citation2011), Ahn et al. (Citation2012), Yacim and Boshoff (Citation2018), and Hastie et al. (Citation2009) describe the basic idea behind Neural Nets as extracting ‘linear combinations of the inputs as derived features, and then modelling the target as a nonlinear function of these features’. The most basic structure of Neural Nets consists of an input layer, one or more hidden layers, and an output layer. Functions of increasing complexity can be modelled by adding more layers and more units within a layer (Goodfellow et al., Citation2016). The strength of individual connections is indicated by weights. Hidden layers find features within the data and allow the following layers to operate directly on those features rather than the entire dataset. By repeatedly adjusting the weights – the strength of individual connections between units – the error rate is minimised.Footnote28 Thus, a Neural Net aims to minimise the errors, where the errors are considered to be a function of the weights of the network (generally the sum of squared errors or cross-entropy). However, as the global minimum of the error function would likely lead to an overfitted solution, some regularisation – either stopping early or a penalty term – is needed. The penalty term is generally implemented via ‘weight decay’, which is a penalty in L2 norm (Hastie et al., Citation2009).

Here, we implement a simple feed-forwards Neural Net model via the ‘nnet-package’ in R (Ripley, Citation2016). This package fits a single hidden-layer Neural Network with two tuning parameters: the number of units in the hidden layer and weight decay (to avoid over-fitting). We calibrate the model via repeated grid searches on combinations of the tuning parameters.

Appendix C.

Description of data

Table C1. Mean values of characteristics in training sets

Table C2. Mean values of characteristics in testing sets