1 Introduction

Given a behavioural model and an outcome space of possible observations, Selten (1991) distinguishes between three types of theories. A point theory gives a single element of the outcome space and predicts this point as the central tendency of the observations. A distribution theory gives a probability distribution over the outcome space and predicts that observations are independently drawn according to this distribution. Finally, an area theory only predicts that the observed outcomes should lie in a certain subset of the outcome space. For example, a distribution theory could predict that some variable of interest is uniformly distributed on the unit interval. A point theory, on the other hand, would predict that the mean (or median) of the observations is equal to 0.5. Finally, an area theory would predict that the observations lie in the interval \([0,1]\). Given this classification, a distribution theory is more informative than either a point theory or an area theory in the sense that if we know the observations to be uniformly distributed, we also know their central tendency (mean or median) and their area (support).

Many applications in experimental and revealed preference settings fall into the class of area theories. With respect to these theories, models are often evaluated on the basis of two metrics: the hit rate and the area. The hit rate gives the percentage of all observations that fall within the predicted subset of the outcome space. A high hit rate implies that many subjects have made choices that are consistent with the model’s predictions. The hit rate, however, only captures one dimension of the model’s performance. In general, the hit rate of a model will be higher if the model becomes less permissive (i.e. the model imposes weaker restrictions on the observed behaviour). Therefore, for an area theory to be meaningful it is desirable that the empirical test is sufficiently strong. The permissiveness can be measured by the ‘area’ of the test, which gives the relative size of the predicted subset compared to the set of all possible outcomes.Footnote 1

Generally, a favourable hit rate, for a specific behavioural model, provides convincing support for the model only if the associated area is sufficiently small. In practice, however, the two measures are almost always positively correlated, which in fact makes it interesting to define a summarizing measure that combines the two measures of empirical performance into a single metric, a so called measure of predictive success. Selten (1991) argues in favour of the functional specification that determines the predictive success as the difference between the hit rate and the area:

$$\begin{aligned} \text {predictive success} = \text {hit rate} - \text {area}. \end{aligned}$$

This measure of predictive success is frequently used experimental studiesFootnote 2 and has recently been advocated for use with revealed preference tests by Beatty and Crawford (2011).Footnote 3 In revealed preference studies, the area is usually quantified as one minus the Bronars (1987) power, which gives the probability that a randomly generated datasets (obtained from a uniform distribution on the budget hyperplanes) will fail the revealed preference test.

Different area theories (and revealed preference models) can be evaluated on the basis of their predictive success, and models with higher predictive success can be seen as having a better empirical fit. However, when comparing the predictive success between two models, it is not at all obvious how big the difference in predictive success needs to be in order to be ‘significant’. The literature dealing with predictive success measures is silent on this point. The main reason for this is that the theory underlying the predictive success measure is not a stochastic theory: the observations are either inside or outside the predicted set (see Hey (1998) for a discussion). However, by considering the space of all possible observed behaviour as the relevant population, we show that it is nevertheless possible to conduct valid statistical inference. Our paper uses elementary large sample theory to construct asymptotically valid confidence intervals for various predictive success measures. In this way it becomes possible to construct asymptotic valid hypothesis tests to verify whether the predictive success of a model is larger than some benchmark threshold (e.g. zero) or to compare the predictive success between different opposing models.

In the next section, we set out the framework and derive the statistical results. Section 3 contains an empirical illustration of our findings that compares the predictive success of different revealed preference tests for models of intertemporal decision making.

2 Framework

The building blocks of our framework are data sets, denoted by \(s\). A dataset may correspond to the outcome of an experiment for a single subject. We denote by \(\Omega \) the set of all possible data sets that can be observed. An experiment is given by a finite number of datasets \(\{s_i\}_{i\le n}\) from \(\Omega \).

2.1 Hit rate

An area theory for a certain model of behaviour predicts that the datasets will fall within a certain subset \(A\) of the outcome space \(\Omega \). Given such area theory, we consider the indicator function \(I:\Omega \rightarrow \{0,1\}:s\mapsto I(s)\) such that \(I(s)=1\) if and only if \(s \in A\). The hit rate, \(r_n\), of the experiment \(\{s_i\}_{i\le n}\) is given by the proportion of datasets that fall within the set \(A\).

$$\begin{aligned} r_n = \frac{1}{n}\sum _{i=1}^n I(s_i). \end{aligned}$$

2.2 Area

In order to define the area, we need a bit more work. To start, let us fix a dataset \(s_i \in \Omega \) and consider a probably space \((\Omega _i, \mathcal {B}_i, \mathbb {F}_i)\) which may depend on the specificities of the dataset \(s_i\). Here, \(\Omega _i \subseteq \Omega \) is a subset of the outcome space such that \(s_i \in \Omega _i\). The set \(\mathcal {B}_i\) is a sigma algebra on \(\Omega _i\) such that the function \(I(.)\) restricted to \(\Omega _i\) is measurable and \(\mathbb {F}_i:\mathcal {B}_i \rightarrow [0,1]\) is a probability measure. We define the area of the dataset \(s_i\) by the function \(\rho (s_i): \Omega \rightarrow [0,1]\) where

$$\begin{aligned} \rho (s_i) = \int I(s) \,\mathbb {F}_i(\text {d}s). \end{aligned}$$

Intuitively, \(\rho (s_i)\) measures the size of the set \(A\) according to the measure \(\mathbb {F}_i\). The area of the experiment \(\{s_i\}_{i\le n}\) is defined as the mean of the areas of the datasets in the experiment:

$$\begin{aligned} a_n = \frac{1}{n}\sum _{i=1}^n \rho (s_i) \end{aligned}$$

In many experimental settings we have that \(\Omega \) is finite, \(\Omega _i = \Omega \) and \(\mathbb {F}_i\) equals the uniform distribution on \(\Omega \), i.e. each individual dataset is given an equal probability. In such setting, \(\rho (s_i)\) will be the same for all \(s_i\) and the measure \(a_n\) will coincide with \(\rho \). Observe, however, that our framework is flexible enough for other specifications of the probability measure \(\mathbb {F}_i\).Footnote 4

In some cases, it is possible to obtain \(\rho (.)\) as a closed form solution. In other settings (like revealed preference theory) no closed form solutions are known. To encompass those situations, we allow \(\rho (s_i)\) to be approximated by simulation. In such cases, we draw \(m\) i.i.d. datasets \(\{\tilde{s}_1^i,\ldots , \tilde{s}_m^i\}\) using the probability measure \(\mathbb {F}_i\) and compute the finite sample approximation:

$$\begin{aligned} \rho _{m}(s_i) = \frac{1}{m}\sum _{k=1}^m I(\tilde{s}_k^i). \end{aligned}$$

The area of the experiment is then approximated by

$$\begin{aligned} a_{n,m} = \frac{1}{n}\sum _{i=1}^n \rho _m(s_i), \end{aligned}$$

Using the law of large numbers, we have that for \(m \rightarrow \infty \)\(a_{n,m} \rightarrow ^{P}\!a_n\).

2.3 Predictive success

The hit rate \(r_n\) and the area \(a_{n,m}\) can be combined in a measure of predictive success \(p:[0,1]^2 \rightarrow \mathbb {R}: (r,a)\mapsto p(r, a)\). Intuitively, \(p(r_n,a_{n,m})\) measures the performance of the behavioural model underlying the indicator function \(I(.)\). Usually, \(p\) is increasing in its first argument and decreasing in its second. We assume that \(p(.,.)\) is continuously differentiable.

2.4 Large sample results

We consider the probability space \((\Omega , \mathcal {B}, \mathbb {P})\) where \(\mathcal {B}\) is a sigma algebra on \(\Omega \) and \(\mathbb {P}\) is a probability distribution on \(\Omega \) giving the law by which the individual datasets in the experiment are obtained. We assume that \(\mathcal {B}\) is such that both the functions \(I(.)\) and \(\rho (.)\) are measurable.

The population hit rate and area are given by

$$\begin{aligned} r = \int I(s) \mathbb {P}(\text {d}s),\qquad \text { and }\qquad a = \int \rho (s) \mathbb {P}(\text {d}s). \end{aligned}$$

Consider an experiment \(\{s_1, \ldots , s_n \}\) which is obtained from \(n\) i.i.d. draws according to the law \(\mathbb {P}\). By the law of large numbers, we have that, as \(n \rightarrow \infty \) and \(m\, n^{-1} \rightarrow \infty \): \(r_n \rightarrow ^{P}\!r\) and \(a_{n,m} \rightarrow ^{P}\!a\). Further, using the classical central limit theorem, we have that

$$\begin{aligned} \sqrt{n}\left( \begin{array}{l}r_n - r\\ a_{n,m}-a\end{array}\right) \rightarrow N\left( 0, \Sigma \right) , \end{aligned}$$

where

$$\begin{aligned} \Sigma = \left[ \begin{array}{l@{\quad }l} r(1-r) &{} \int (I(s)-r)(\rho (s)-a){\mathbb {P}}(\text {d}s)\\ \ \int (I(s)-r)(\rho (s)-a)\mathbb {P}(\text {d}s) &{} \int (\rho (s)-a)^2\mathbb {P}(\text {d}s) \end{array}\right] , \end{aligned}$$

is the asymptotic variance–covariance matrix. The elements of \(\Sigma \) can be consistently estimated by their finite sample analogues.

$$\begin{aligned} S_{n,m} = \left[ \begin{array}{l@{\quad }l} r_n(1-r_n) &{} \frac{1}{n}\sum _i (I(s_i)-r_n)(\rho _{m}(s_i)-a_{n,m})\\ \frac{1}{n}\sum _i (I(s_i)-r_n)(\rho _{m}(s_i)-a_{n,m}) &{} \frac{1}{n}\sum (\rho _{m}(s_i)-a_{n,m})^2 \end{array}\right] . \end{aligned}$$

Using the continuous mapping theorem, we have that for \(n \rightarrow \infty \) and \(m\,n^{-1}\rightarrow \infty \): \(p(r_n, a_{n,m}) \rightarrow ^P p(r,a)\). Next, let \(\delta \) be the row vector of partial derivatives of the predictive success measure \(p(r,a)\) evaluated at \((r,a)\),

$$\begin{aligned} \delta = \left[ \begin{array}{c@{\quad }c} \frac{\partial p(r,a)}{\partial r}&\frac{\partial p(r,a)}{\partial a} \end{array} \right] . \end{aligned}$$

Using the delta method, we obtain that, for \(n \rightarrow \infty \) and \(m\, n^{-1} \rightarrow \infty \),

$$\begin{aligned} \sqrt{n}\left( p(r_n,a_{n,m}) - p(r,a)\right) \rightarrow N\left( 0, \delta \Sigma \delta ^{\prime }\right) . \end{aligned}$$

The variance, \(\delta \Sigma \delta ^\prime \), can be consistently estimated by

$$\begin{aligned} v_{n,m} = \delta _{n,m} S_{n,m} \delta _{n,m}^\prime , \end{aligned}$$

where

$$\begin{aligned} \delta _{n,m} = \left[ \begin{array}{c@{\quad }c} \frac{\partial p(r_n,a_{n,m})}{\partial r}&\frac{\partial p(r_n,a_{n,m})}{\partial a}\end{array}\right] . \end{aligned}$$

If \(\Phi (.)\) is the standard normal cdf function, and \(c_{a}\) is defined by,

$$\begin{aligned} \Phi (c_\alpha ) - \Phi (-c_\alpha ) = \alpha , \end{aligned}$$

then

$$\begin{aligned} C_{n,m}^\alpha = \left[ \begin{array}{ll}p(r_n,a_{n,m})-c_\alpha \sqrt{\frac{v_{n,m}}{n}},&p(r_n,a_{n,m})+c_\alpha \sqrt{\frac{v_{n,m}}{n}}\end{array}\right] \end{aligned}$$

is an asymptotic \(\alpha \times 100\) % confidence interval for the predictive success measure \(p(r,a)\).

2.5 Comparing predictive success

In many cases it is also interesting to compare two tests on the basis of their difference in predictive success. Consider two tests with hit rates and area equal to \(r, a\) and \(\tilde{r}, \tilde{a}\), respectively. By the central limit theorem, we know that,

$$\begin{aligned} \sqrt{n} \left( \begin{array}{l}r_n - r\\ a_{n,m}-a\\ \tilde{r}_n -\tilde{r}\\ \tilde{a}_{n,m} - \tilde{a} \end{array} \right) \rightarrow N\left( 0, \Sigma _{\Delta }\right) , \end{aligned}$$

where \(\Sigma _{\Delta }\) is the asymptotic variance covariance matrix whose elements can be consistently estimated using the finite sample plug-ins. For example, the covariance between \(r\) and \(\tilde{r}\) is equal to

$$\begin{aligned} \int (I(s)-r)(\tilde{I}(s)-\tilde{r})\mathbb {P}(\text {d}s), \end{aligned}$$

which can be consistently estimated by

$$\begin{aligned} \frac{1}{n} \sum _i (I(s_i)-r_n)(\tilde{I}(s_i)-\tilde{r}_n). \end{aligned}$$

We denote the estimator of the variance–covariance matrix by \(S_{\Delta ,n,m}\). Again, using the delta method, the asymptotic distribution of the difference in predictive success is given by

$$\begin{aligned} \sqrt{n}\left[ \left( p(r_n,a_{n,m})\quad -p(\tilde{r}_n, \tilde{a}_{n,m})\right) - \left( p(r,a) - p(\tilde{r}, \tilde{a})\right) \right] \rightarrow N\left( 0, \delta _{\Delta } \Sigma _{\Delta } \delta _{\Delta }^{\prime }\right) , \end{aligned}$$

where \(\delta _{\Delta }\) is equal to the following row vector of partial derivatives:

$$\begin{aligned} \delta _{\Delta } = \left[ \begin{array}{cccc} \frac{\partial p(r,a)}{\partial r}&\frac{\partial p(r,a)}{\partial a}&-\frac{\partial p(\tilde{r}, \tilde{a})}{\partial r}&-\frac{\partial p(\tilde{r}, \tilde{a})}{\partial a} \end{array} \right] . \end{aligned}$$

Set

$$\begin{aligned} v_{\Delta ,n,m} = \delta _{\Delta , n,m} S_{\Delta ,n,m} \delta _{\Delta ,n,m}^{\prime }, \end{aligned}$$

where

$$\begin{aligned} \delta _{\Delta ,n,m} = \left[ \begin{array}{cccc}\frac{\partial p(r_n,a_{n,m})}{\partial r}&\frac{\partial p(r_n,a_{n,m})}{\partial a}&-\frac{\partial p(\tilde{r}_n, \tilde{a}_{n,m})}{\partial r}&-\frac{\partial p(\tilde{r}_n, \tilde{a}_{n,m})}{\partial a}\end{array}\right] . \end{aligned}$$

Then

$$\begin{aligned} \left[ \begin{array}{ll}p(r_n,a_{n,m})-p(\tilde{r}_n, \tilde{a}_{n,m})-c_\alpha \sqrt{\frac{v_{\Delta , n,m}}{n}},&p(r_n,a_{n,m})-p(\tilde{r}_n,\tilde{a}_{n,m})+c_\alpha \sqrt{\frac{v_{\Delta ,n,m}}{n}}\end{array}\right] \end{aligned}$$

is an asymptotic \(\alpha \times 100\) % CI for \(p(r,a)-p(\tilde{r},\tilde{a})\).

3 Illustration

We illustrate our results using various revealed preference tests for different models of intertemporal decision making. The first model is the standard life cycle (LC) model where an individual optimizes a time separable additive utility function \(\sum _t \delta ^t u(\mathbf {q}_t)\) subject to an intertemporal budget constraint \(\mathbf {p}_t \mathbf {q}_t + a_t = I_t + (1+r_t)a_{t-1}\). Here \(\delta < 1\) is a subjective discount rate, \(\mathbf {p}_t\) are the period \(t\) prices, \(a_t\) is the value of assets at period \(t\), \(I_t\) is the contemporaneous income and \(r_t\) is the interest rate. Datasets for this model are determined by prices, quantities and interest rates for a finite number of periods, \(s_i = \{\mathbf {p}_{t,i}, \mathbf {q}_{t,i}, r_{t,i}\}_{t = 1,\ldots |T|}\). The revealed preference conditions for this life cycle model were derived by Browning (1989).

For the second model, let us first single out a habit forming good \(c\). The habits (H) model replaces the intertemporal separable utility function by a utility function of the form \(\sum _t \delta ^t u(\mathbf {q}_t, c_{t-1})\). Here, the consumption of the addictive good in period \(t-1\) is allowed to influence the utility in period \(t\). The revealed preference characterization of this model was given by Crawford (2010).

Our third model, the habits as durables (HAD) model, considers a variant where the intertemporal utility function is given by \(\sum _t \delta ^t u(\mathbf {q}_t, A_t)\) and where \(A_t = \beta A_{t-1} + c_t\) represents a stock of addiction with depreciation rate \(\beta \) that determines how fast the addiction wears off. This is the rational addiction model put forward by Becker and Murphy (1988). The revealed preference characterization of this model was derived by Demuynck and Verriest (2013).

As a final fourth model, we consider the static utility maximization model where the household maximizes each period a time-independent utility function \(u(\mathbf {q})\) subject to a budget constraint \(\mathbf {p}_t \mathbf {q}= m_t\) for some level of expenditure \(m_t\). The revealed preference conditions for this model are given by the Generalized Axiom of Revealed Preference (GARP) (see, for example, Varian (1982)). Typically, in a revealed preference setting, we specify the measure \(\mathbb {F}_i\) as the probability law that randomly samples datasets \(\tilde{s}_i = \{\mathbf {p}_t^i, \tilde{\mathbf {q}}_t^i\}_{t \in T}\) where \(\tilde{\mathbf {q}}_t^i\) is obtained by a uniform draw from the hyperplane \(\{\mathbf {q}\in \mathbb {R}^n_+|\mathbf {p}_t^i \mathbf {q}= \mathbf {p}_t^i \mathbf {q}_t^i\}\). This is analogue to the way that the Bronars (1987) power is computed.

We consider 3 measures of predictive success.

$$\begin{aligned} p_1(r,a)&= r-a,\\ p_2(r,a)&= \frac{r}{a},\\ p_3(r,a)&= \frac{r-a}{1-a}. \end{aligned}$$

The first measure takes the difference between the hit rate and the area and is the measure that has become standard in the literature. It is bounded between -1 and 1. In the best case scenario, \(r \rightarrow 1\) and \(a\rightarrow 0\). This gives a predictive success close to one. In such case, most datasets pass the test while the area is very small. In the worst case scenario, \(r\rightarrow 0\) and \(a\rightarrow 1\), which give a predictive success close to minus one. In this case, almost all observations are inconsistent with the model while the area is almost equal to the outcome space \(\Omega \). In intermediate cases, the measure of predictive success is found somewhere between minus one and plus one. Zero is a natural benchmark where \(r = a\).

The second measure takes the ratio of the hit rate and area. Intuitively, \(p_2\) measures the density of the observed datasets within the predicted area. It is bounded from below by zero. The natural benchmark, where \(r = a\), gives a predictive success equal to one. The third measure is obtained from the first measure by dividing it by the maximal value that it can obtain for fixed \(a\). It can also be written as \(1-\frac{1-r}{1-a}\). Intuitively, the higher the predictive success measure will be, the lower the density outside the predicted area. Its benchmark is equal to zero. We refer to Selten (1991) for a more thorough discussion of the differences between these predictive success measures.

3.1 Data description

We use data from the Encuesta Continua de Presupuestos Familiares. This dataset contains detailed information on consumed quantities and prices for a large sample of Spanish households. We refer to Browning and Collado (2001), Crawford (2010) and Demuynck and Verriest (2013) for a more detailed explanation of this data set. The observations range from 1985 to 1997 and are obtained on a quarterly basis. Every quarter, new households are participating in the moving panel and others are dropped. There are a maximum of eight consecutive observations per household. We consider 14 nondurable commodity categoriesFootnote 5 and take tobacco as the habit forming good.Footnote 6 We have a sample of \(671\) households (\(n = 671\)). Finally, we simulate the areas \(\rho (s_i)\) using 1,000 random draws per dataset (in other words, we set \(m\) equal to 1,000).

3.2 Results

Table 1 provides the results on the estimates of \(p(r,a)\) for the different measures and the 95 % asymptotic confidence intervals. For the first measure, the highest estimate is for the HAD model which is also the only model whose confidence interval excludes the benchmark value 0. For the second measure, the highest value is found for the LC model. However, this model also has the highest variance, which makes its value highly uncertain. Both H and HAD models exclude 1 from the 95 % confidence intervals. The last measure gives qualitatively similar results as the first measure.

Table 1 Mean values, sample standard deviations \((\sqrt{v_{n,m}})\) and 95 % confidence intervals for the predictive success measures

Table 2 gives the mean values and 95 % asymptotic confidence intervals for the difference in predictive success between the different revealed preference tests. Many intervals include the value of zero meaning that the hypothesis of equal predictive success cannot be rejected at the 5 % level. Exceptions to this are the differences between the GARP and the HAD test for measures 1 and 2, the difference between the LC and HAD test for measures 1 and 3 and the difference between the H and HAD test for all predictive success measures under consideration.

Table 2 Mean, sample standard deviation \((\sqrt{v_{\Delta ,n,m}})\) and 95 % confidence intervals for difference in predictive success

3.3 Size analysis

Our results are based on large sample statistics. This means that they may be unreliable if the number of datasets in the experiment is small. In order to analyse this, we conduct a simple level analysis based on an artificial dataset on 10 observations (\(|T| = 10\)) and 10 goods.Footnote 7 We compute the area of this dataset using the Bronars procedure. We consider the case where the null-hypotheses \(p_1(r,a) = 0, p_2(r,a) = 1\) and \(p_3(r,a) = 0\) hold. Towards this end, we randomly generated experiments of various sizes using the same Bronars procedure.Footnote 8 Table 3 gives the level of the test that rejects if the sample-based predictive success measure falls outside the 95 % confidence interval. The table shows that the test coincides with the nominal level for experiments of 250 datasets or more. Small experiments, however, tend to reject the null-hypothesis to often.

Table 3 Level of the test that rejects the hypothesis \(p_1(r,a) = 0; p_2(r,a) = 1\) or \(p_3(r,a)=0\) if these values fall outside the 95 % confidence intervals

4 Conclusion

This note provides statistical inference for measures of predictive success. Predictive success measures are frequently used to evaluate and compare the performance of different models of individual and group behaviour in experimental and revealed preference studies. Our results allow us to derive confidence intervals for the value of a predictive success measure or for the difference between the predictive success measure of two opposing models. We provide a brief illustration of our findings by comparing the predictive success of different revealed preference tests for models of intertemporal decision making. Finally, simulation results indicate that our tests give reliable results for moderately sized experiments but that type I errors may be above the 5 % nominal value in small samples.