Abstract
We compare different procedures which generate \(D_{B}\)-efficient designs for choice-based conjoint analysis using the latent class mixed logit model which captures latent consumer heterogeneity in a flexible way. These procedures are the Coordinate-Exchange algorithm, the Relabel-Swap-Cycle algorithm, and the remaining six combinations of the individual components of the latter. Halton draws and a minimum potential design for prior draws both of which reduce computation times serve to determine \(D_{B}\)-errors of designs. We simulate choices for each set of generated designs and constellations which differ with respect to number of choice sets, number of clusters, within cluster heterogeneity, amount of stochastic error, relative cluster size and cluster similarity. Using these artificial choices we estimate parameters of the latent class mixed logit model in the next step. Designs are evaluated by TOPSIS scores which combine estimation accuracy and run time. ANOVA with TOPSIS scores as dependent variable shows that Relabel alone yields the best results of all procedures investigated. Coordinate-Exchange, Swap alone and the combination of Relabel and Swap turn out to be second best. Relabel also leads to much lower run times than the other procedures. We recommend to use Relabel and to avoid Cycle altogether because it performs worst.
Similar content being viewed by others
References
Andrews RL, Ansari A, Currim IS (2002) Hierarchical bayes versus finite mixture conjoint analysis models: a comparison of fit, prediction and partworth recovery. J Mark Res 39:87–98
Atkinson AC, Donev AN (1992) Optimum experimental designs. Oxford University Press, Clarendon
Baier T, Neuwirth E (2007) Excel : COM : R. Comput Stat 22:91–108
Bhat CR (2001) Quasi-random maximum simulated likelihood estimation of the mixed multinomial logit model. Transp Res Part B 35:677–693
Bliemer MCJ, Rose JM (2010) Construction of experimental designs for mixed logit models allowing for correlation across choice observations. Transp Res Part B 44:720–734
Chaloner K, Verdinelli I (1995) Bayesian experimental design: a review. Stat Sci 10:273–304
Dhar R, Simonson I (2003) The effect of forced choice on choice. J Mark Res 40:146–160
Evgeniou T, Pontil M, Toubia O (2007) A convex optimization approach to modeling consumer heterogeneity in conjoint estimation. Mark Sci 26:805–818
Ferrini S, Scarpa R (2007) Designs with a-priori information for nonmarket valuation with choice-experiments: a monte carlo study. J Environ Econ Manage 53:342–363
Greene WH (1997) Economic analysis. Prentice Hall, Upper Saddle River
Greene WH, Hensher DA (2013) Revealing additional dimensions of preference heterogeneity in a latent class mixed multinomial logit model. Appl Econ 45:1897–1902
Gustafsson A, Herrmann A, Huber F (2007) Conjoint analysis as an instrument of market research practice. In: Gustafsson A, Herrmann A, Huber F (eds) Conjoint measurement: methods and application. Springer, Berlin, pp 3–30
Huber J, Zwerina K (1996) The importance of utility balance in efficient choice designs. J Mark Res 33:307–317
Kanninen B (2002) Optimal design for multinomial choice experiments. J Mark Res 39:214–222
Kessels R, Goos P, Vandebroek M (2004) Comparing algorithms and criteria for designing Bayesian conjoint choice experiments. DTEW Research Report 0427, pp 1–38
Kessels R, Goos P, Vandebroek M (2006) A comparison of criteria to design efficiency choice experiments. J Mark Res 43:409–419
Kessels R, Jones B, Goos P, Vandebroek M (2009) An efficient algorithm for constructing Bayesian optimal choice designs. J Bus Econ Stat 27:279–291
McFadden D (1984) Econometric analysis of qualitative response models. In: Griliches Z, Intriligator MD (eds) Handbook of econometrics, vol 2. North Holland, New York, pp 1395–1457
McFadden D, Train KE (2000) Mixed MNL models for discrete response. J Appl Econ 15:447–470
Nychka D, Furrer R, Paige J, Sain S (2015) Fields: tools for spatial data. R Package Version 9.0. http://CRAN.R-project.org/package=fields. Accessed 9 May 2016
Rose JM, Scarpa R, Bliemer MCJ (2009) Incorporating model uncertainty into the generation of efficient stated choice experiments: a model averaging approach. Working Paper, Institute of Transport and Logistics Studies, University of Sydney
Sándor Z, Wedel M (2001) Designing conjoint choice experiments using managers’ prior beliefs. J Mark Res 38:430–444
Sándor Z, Wedel M (2002) Profile construction in experimental choice designs for mixed logit models. Mark Sci 21:455–475
Scarpa R, Rose JM (2008) Design efficiency for non-market valuation with choice modeling: how to measure it, what to report and why. Aust J Agric Resour Econ 52:253–282
Street DJ, Burgess L, Louviere JJ (2005) Quick and easy choice sets: constructing optimal and nearly optimal stated choice experiments. Int J Res Mark 22:459–470
Toubia O, Hauser JR, Simester DI (2004) Polyhedral methods for adaptive choice-based conjoint analysis. J Mark Res 41:116–131
Train KE (2000) Halton sequences for mixed logit. Working Paper, Department of Economics, University of California at Berkeley
Train KE (2003) Discrete choice methods with simulation. Cambridge University Press, New York
Train KE (2008) EM algorithms for nonparametric estimation of mixing distributions. J Choice Model 1:40–69
Vermeulen B, Goos P, Scarpa R, Vandebroek M (2011) Bayesian conjoint choice designs for measuring willingness to pay. Environ Resour Econ 48:129–149
Vriens M, Wedel M, Wilms T (1996) Metric conjoint segmentation methods: a monte carlo comparison. J Mark Res 33:73–85
Yoon KP, Hwang C (1995) Multiple attribute decision making: an introduction. SAGE publications, California
Yu J, Goos P, Vandebroek M (2009) Efficient conjoint choice designs in the presence of respondent heterogeneity. Mark Sci 28:122–135
Acknowledgements
We thank two anonymous reviewers for their useful comments which helped us to improve the paper.
Author information
Authors and Affiliations
Corresponding author
Appendices
Appendix A: Model estimation
We set initial cluster shares \(g_{c}^{0}\) to equal values, initial cluster means \(b_{c}^{0}\) to a random value in the interval [\(\beta - 2 \cdot \sigma , \beta + 2 \cdot \sigma \)] with \(\beta \) being the true value of the cluster means and \(\sigma =0.1\), which corresponds to the higher value of factor 3. The initial covariance matrix is diagonal with all diagonal elements equaling 0.5. We decide to initialize parameters this way, because we focus on the comparison of design-generating procedures and do not intend to evaluate estimation algorithms.
The proper estimation algorithm is taken from Train (2008):
-
1.
For each subject R random values are drawn from the cluster-specific normal distribution \(N(b_{c}^{0},V_{c}^{0})\). The r-th draw for subject n in cluster c is labeled as \({\hat{\beta }}_{ncr}^{0}\). A preliminary study has shown that four draws (\(R=4\)) are sufficient.
-
2.
For each subject in each cluster and each draw a weight is calculated as
$$\begin{aligned} h_{ncr}=\frac{g_{c}^{0} K_{n}\left( {\hat{\beta }}_{ncr}^{0}\right) }{\sum _{c} g_{c}^{0} \sum _{r}K_{n}({\hat{\beta }}_{ncr}^{0}/R} \end{aligned}$$with \(K_n(\hat{\beta }_{ncr}^0)\) as probability of the choices \(y_n\) conditional on \(\hat{\beta }_{ncr}^0\) (\(y_{nt}\) is the index of the alternative chosen from choice set t)
$$\begin{aligned} K_{n}({\hat{\beta }}_{ncr}^{0})=\prod _{t}{\frac{e^{x'_{ny_{nt}t} {\hat{\beta }}_{ncr}^{0}}}{\sum _j{e^{x'_{njt}{\hat{\beta }}_{ncr}^{0}}}}} \end{aligned}$$ -
3.
Cluster means, covariances and shares are updated as follows:
$$\begin{aligned} b_{c}^{1}= & {} \frac{\sum _{n}\sum _{r} h_{ncr}{\hat{\beta }}_{ncr}^{0}}{\sum _{n}\sum _{r}h_{ncr}}\\ V_{c}^{1}= & {} \frac{\sum _{n}\sum _{r} h_{ncr} [({\hat{\beta }}^{0}_{ncr} -b_{c}^{1})({\hat{\beta }}_{ncr}^{0}-b_{c}^{1})']}{\sum _{n}\sum _{r}h_{ncr}},\\ g_{c}^{1}= & {} \frac{\sum _{n}\sum _{r}h_{ncr}}{\sum _{n}\sum _{c'} \sum _{r}h_{nc'r}} \end{aligned}$$ -
4.
If the absolute difference between all previous and updated parameter values is less than \(10^{-6}\) or 10000 iterations have been made, the algorithm stops; otherwise, it sets the old parameter values to their updated values and goes back to step 1.
We need individual coefficients \({\hat{\beta }}_{n}\) to compute the estimation accuracy defined in expression (14). Individual coefficients are a linear combination of cluster specific mean coefficients weighted by cluster membership probabilities of the respective subject \({\hat{\beta }}_{n} = \sum _{c} u_{nc} \, b_{c}\), where \(u_{nc}\) is the cluster membership probability of subject n to cluster c:
Appendix B: TOPSIS
Let Z be an \((m \times n)\) matrix where m is the number of values, n the number of criteria and let \(z_{ij}\) denote an element of this matrix.
-
1.
Normalize to get
$$\begin{aligned} r_{ij} =\frac{z_{ij}}{\sqrt{\sum _{i=1}^{m} z_{ij}^{2}}} \end{aligned}$$ -
2.
Calculate weighted normalized values (we set the weight for estimation accuracy to 0.67 and that for run time to 0.33):
$$\begin{aligned} t_{ij} = w_{j}r_{ij} \end{aligned}$$ -
3.
Determine the best and worst value \(t_{bj}\) and \(t_{wj}\) for every criterion j.
-
4.
Calculate the distance of every value i to the best and worst values:
$$\begin{aligned} d_{ib} = \sqrt{\sum _{j=1}^{n}(t_{ij}-t_{bj})^{2}}, \quad d_{iw} = \sqrt{\sum _{j=1}^{n}(t_{ij}-t_{wj})^{2}} \end{aligned}$$ -
5.
Calculate for each value i its similarity to the best values:
$$\begin{aligned} s_{i}=\frac{d_{iw}}{d_{ib}+d_{iw}} \end{aligned}$$
\(s_{i}\) lies in the interval [0, 1] with 0 marking the worst value and 1 marking the best value.
Rights and permissions
About this article
Cite this article
Falke, A., Hruschka, H. A Monte Carlo study of design-generating algorithms for the latent class mixed logit model. OR Spectrum 39, 1035–1053 (2017). https://doi.org/10.1007/s00291-017-0483-1
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00291-017-0483-1