Skip to main content
Log in

A Monte Carlo study of design-generating algorithms for the latent class mixed logit model

  • Regular Article
  • Published:
OR Spectrum Aims and scope Submit manuscript

Abstract

We compare different procedures which generate \(D_{B}\)-efficient designs for choice-based conjoint analysis using the latent class mixed logit model which captures latent consumer heterogeneity in a flexible way. These procedures are the Coordinate-Exchange algorithm, the Relabel-Swap-Cycle algorithm, and the remaining six combinations of the individual components of the latter. Halton draws and a minimum potential design for prior draws both of which reduce computation times serve to determine \(D_{B}\)-errors of designs. We simulate choices for each set of generated designs and constellations which differ with respect to number of choice sets, number of clusters, within cluster heterogeneity, amount of stochastic error, relative cluster size and cluster similarity. Using these artificial choices we estimate parameters of the latent class mixed logit model in the next step. Designs are evaluated by TOPSIS scores which combine estimation accuracy and run time. ANOVA with TOPSIS scores as dependent variable shows that Relabel alone yields the best results of all procedures investigated. Coordinate-Exchange, Swap alone and the combination of Relabel and Swap turn out to be second best. Relabel also leads to much lower run times than the other procedures. We recommend to use Relabel and to avoid Cycle altogether because it performs worst.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Similar content being viewed by others

References

  • Andrews RL, Ansari A, Currim IS (2002) Hierarchical bayes versus finite mixture conjoint analysis models: a comparison of fit, prediction and partworth recovery. J Mark Res 39:87–98

    Article  Google Scholar 

  • Atkinson AC, Donev AN (1992) Optimum experimental designs. Oxford University Press, Clarendon

    Google Scholar 

  • Baier T, Neuwirth E (2007) Excel : COM : R. Comput Stat 22:91–108

    Article  Google Scholar 

  • Bhat CR (2001) Quasi-random maximum simulated likelihood estimation of the mixed multinomial logit model. Transp Res Part B 35:677–693

    Article  Google Scholar 

  • Bliemer MCJ, Rose JM (2010) Construction of experimental designs for mixed logit models allowing for correlation across choice observations. Transp Res Part B 44:720–734

    Article  Google Scholar 

  • Chaloner K, Verdinelli I (1995) Bayesian experimental design: a review. Stat Sci 10:273–304

    Article  Google Scholar 

  • Dhar R, Simonson I (2003) The effect of forced choice on choice. J Mark Res 40:146–160

    Article  Google Scholar 

  • Evgeniou T, Pontil M, Toubia O (2007) A convex optimization approach to modeling consumer heterogeneity in conjoint estimation. Mark Sci 26:805–818

    Article  Google Scholar 

  • Ferrini S, Scarpa R (2007) Designs with a-priori information for nonmarket valuation with choice-experiments: a monte carlo study. J Environ Econ Manage 53:342–363

    Article  Google Scholar 

  • Greene WH (1997) Economic analysis. Prentice Hall, Upper Saddle River

    Google Scholar 

  • Greene WH, Hensher DA (2013) Revealing additional dimensions of preference heterogeneity in a latent class mixed multinomial logit model. Appl Econ 45:1897–1902

    Article  Google Scholar 

  • Gustafsson A, Herrmann A, Huber F (2007) Conjoint analysis as an instrument of market research practice. In: Gustafsson A, Herrmann A, Huber F (eds) Conjoint measurement: methods and application. Springer, Berlin, pp 3–30

    Chapter  Google Scholar 

  • Huber J, Zwerina K (1996) The importance of utility balance in efficient choice designs. J Mark Res 33:307–317

    Article  Google Scholar 

  • Kanninen B (2002) Optimal design for multinomial choice experiments. J Mark Res 39:214–222

    Article  Google Scholar 

  • Kessels R, Goos P, Vandebroek M (2004) Comparing algorithms and criteria for designing Bayesian conjoint choice experiments. DTEW Research Report 0427, pp 1–38

  • Kessels R, Goos P, Vandebroek M (2006) A comparison of criteria to design efficiency choice experiments. J Mark Res 43:409–419

    Article  Google Scholar 

  • Kessels R, Jones B, Goos P, Vandebroek M (2009) An efficient algorithm for constructing Bayesian optimal choice designs. J Bus Econ Stat 27:279–291

    Article  Google Scholar 

  • McFadden D (1984) Econometric analysis of qualitative response models. In: Griliches Z, Intriligator MD (eds) Handbook of econometrics, vol 2. North Holland, New York, pp 1395–1457

    Google Scholar 

  • McFadden D, Train KE (2000) Mixed MNL models for discrete response. J Appl Econ 15:447–470

    Article  Google Scholar 

  • Nychka D, Furrer R, Paige J, Sain S (2015) Fields: tools for spatial data. R Package Version 9.0. http://CRAN.R-project.org/package=fields. Accessed 9 May 2016

  • Rose JM, Scarpa R, Bliemer MCJ (2009) Incorporating model uncertainty into the generation of efficient stated choice experiments: a model averaging approach. Working Paper, Institute of Transport and Logistics Studies, University of Sydney

  • Sándor Z, Wedel M (2001) Designing conjoint choice experiments using managers’ prior beliefs. J Mark Res 38:430–444

    Article  Google Scholar 

  • Sándor Z, Wedel M (2002) Profile construction in experimental choice designs for mixed logit models. Mark Sci 21:455–475

    Article  Google Scholar 

  • Scarpa R, Rose JM (2008) Design efficiency for non-market valuation with choice modeling: how to measure it, what to report and why. Aust J Agric Resour Econ 52:253–282

    Article  Google Scholar 

  • Street DJ, Burgess L, Louviere JJ (2005) Quick and easy choice sets: constructing optimal and nearly optimal stated choice experiments. Int J Res Mark 22:459–470

    Article  Google Scholar 

  • Toubia O, Hauser JR, Simester DI (2004) Polyhedral methods for adaptive choice-based conjoint analysis. J Mark Res 41:116–131

    Article  Google Scholar 

  • Train KE (2000) Halton sequences for mixed logit. Working Paper, Department of Economics, University of California at Berkeley

  • Train KE (2003) Discrete choice methods with simulation. Cambridge University Press, New York

    Book  Google Scholar 

  • Train KE (2008) EM algorithms for nonparametric estimation of mixing distributions. J Choice Model 1:40–69

    Article  Google Scholar 

  • Vermeulen B, Goos P, Scarpa R, Vandebroek M (2011) Bayesian conjoint choice designs for measuring willingness to pay. Environ Resour Econ 48:129–149

    Article  Google Scholar 

  • Vriens M, Wedel M, Wilms T (1996) Metric conjoint segmentation methods: a monte carlo comparison. J Mark Res 33:73–85

    Article  Google Scholar 

  • Yoon KP, Hwang C (1995) Multiple attribute decision making: an introduction. SAGE publications, California

    Book  Google Scholar 

  • Yu J, Goos P, Vandebroek M (2009) Efficient conjoint choice designs in the presence of respondent heterogeneity. Mark Sci 28:122–135

    Article  Google Scholar 

Download references

Acknowledgements

We thank two anonymous reviewers for their useful comments which helped us to improve the paper.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Andreas Falke.

Appendices

Appendix A: Model estimation

We set initial cluster shares \(g_{c}^{0}\) to equal values, initial cluster means \(b_{c}^{0}\) to a random value in the interval [\(\beta - 2 \cdot \sigma , \beta + 2 \cdot \sigma \)] with \(\beta \) being the true value of the cluster means and \(\sigma =0.1\), which corresponds to the higher value of factor 3. The initial covariance matrix is diagonal with all diagonal elements equaling 0.5. We decide to initialize parameters this way, because we focus on the comparison of design-generating procedures and do not intend to evaluate estimation algorithms.

The proper estimation algorithm is taken from Train (2008):

  1. 1.

    For each subject R random values are drawn from the cluster-specific normal distribution \(N(b_{c}^{0},V_{c}^{0})\). The r-th draw for subject n in cluster c is labeled as \({\hat{\beta }}_{ncr}^{0}\). A preliminary study has shown that four draws (\(R=4\)) are sufficient.

  2. 2.

    For each subject in each cluster and each draw a weight is calculated as

    $$\begin{aligned} h_{ncr}=\frac{g_{c}^{0} K_{n}\left( {\hat{\beta }}_{ncr}^{0}\right) }{\sum _{c} g_{c}^{0} \sum _{r}K_{n}({\hat{\beta }}_{ncr}^{0}/R} \end{aligned}$$

    with \(K_n(\hat{\beta }_{ncr}^0)\) as probability of the choices \(y_n\) conditional on \(\hat{\beta }_{ncr}^0\) (\(y_{nt}\) is the index of the alternative chosen from choice set t)

    $$\begin{aligned} K_{n}({\hat{\beta }}_{ncr}^{0})=\prod _{t}{\frac{e^{x'_{ny_{nt}t} {\hat{\beta }}_{ncr}^{0}}}{\sum _j{e^{x'_{njt}{\hat{\beta }}_{ncr}^{0}}}}} \end{aligned}$$
  3. 3.

    Cluster means, covariances and shares are updated as follows:

    $$\begin{aligned} b_{c}^{1}= & {} \frac{\sum _{n}\sum _{r} h_{ncr}{\hat{\beta }}_{ncr}^{0}}{\sum _{n}\sum _{r}h_{ncr}}\\ V_{c}^{1}= & {} \frac{\sum _{n}\sum _{r} h_{ncr} [({\hat{\beta }}^{0}_{ncr} -b_{c}^{1})({\hat{\beta }}_{ncr}^{0}-b_{c}^{1})']}{\sum _{n}\sum _{r}h_{ncr}},\\ g_{c}^{1}= & {} \frac{\sum _{n}\sum _{r}h_{ncr}}{\sum _{n}\sum _{c'} \sum _{r}h_{nc'r}} \end{aligned}$$
  4. 4.

    If the absolute difference between all previous and updated parameter values is less than \(10^{-6}\) or 10000 iterations have been made, the algorithm stops; otherwise, it sets the old parameter values to their updated values and goes back to step 1.

We need individual coefficients \({\hat{\beta }}_{n}\) to compute the estimation accuracy defined in expression (14). Individual coefficients are a linear combination of cluster specific mean coefficients weighted by cluster membership probabilities of the respective subject \({\hat{\beta }}_{n} = \sum _{c} u_{nc} \, b_{c}\), where \(u_{nc}\) is the cluster membership probability of subject n to cluster c:

$$\begin{aligned} u_{nc} =\frac{\sum _{r}h_{ncr}}{\sum _{c'} \sum _{r}h_{nc'r}} \end{aligned}$$

Appendix B: TOPSIS

Let Z be an \((m \times n)\) matrix where m is the number of values, n the number of criteria and let \(z_{ij}\) denote an element of this matrix.

  1. 1.

    Normalize to get

    $$\begin{aligned} r_{ij} =\frac{z_{ij}}{\sqrt{\sum _{i=1}^{m} z_{ij}^{2}}} \end{aligned}$$
  2. 2.

    Calculate weighted normalized values (we set the weight for estimation accuracy to 0.67 and that for run time to 0.33):

    $$\begin{aligned} t_{ij} = w_{j}r_{ij} \end{aligned}$$
  3. 3.

    Determine the best and worst value \(t_{bj}\) and \(t_{wj}\) for every criterion j.

  4. 4.

    Calculate the distance of every value i to the best and worst values:

    $$\begin{aligned} d_{ib} = \sqrt{\sum _{j=1}^{n}(t_{ij}-t_{bj})^{2}}, \quad d_{iw} = \sqrt{\sum _{j=1}^{n}(t_{ij}-t_{wj})^{2}} \end{aligned}$$
  5. 5.

    Calculate for each value i its similarity to the best values:

    $$\begin{aligned} s_{i}=\frac{d_{iw}}{d_{ib}+d_{iw}} \end{aligned}$$

\(s_{i}\) lies in the interval [0, 1] with 0 marking the worst value and 1 marking the best value.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Falke, A., Hruschka, H. A Monte Carlo study of design-generating algorithms for the latent class mixed logit model. OR Spectrum 39, 1035–1053 (2017). https://doi.org/10.1007/s00291-017-0483-1

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00291-017-0483-1

Keywords

Navigation