Skip to main content

Maximum Entropy Design in High Dimensions by Composite Likelihood Modelling

  • Conference paper
  • 756 Accesses

Part of the book series: Contributions to Statistics ((CONTRIB.STAT.))

Abstract

In maximum entropy sampling (MES), a design is chosen by maximizing the joint Shannon entropy of parameters and observations. However, when the conditional parametric model of the response contains a large number of covariates, the posterior calculations in MES can be challenging or infeasible. In this work, we consider the use of composite likelihood modelling to break down the complexity of the full likelihood and code the original optimization problem into a set of simple partial likelihood problems. We study the optimality behaviour of the composite likelihood sampling approach as the number of design variables grows using both asymptotic analysis and numerical simulations.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   129.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  • Cox, D.R., Reid, N.: A note on pseudolikelihood constructed from marginal densities. Biometrika 91, 729–737 (2004)

    Article  MathSciNet  MATH  Google Scholar 

  • Goldberg, D.E.: Genetic Algorithms in Search, Optimization and Machine Learning. Kluwer Academic, Boston (1989)

    MATH  Google Scholar 

  • Johnson, M.E., Moore, L.M., Ylvisaker, D.: Minimax and maximin distance design. J. Stat. Plan. Inference 26, 131–148 (1990)

    Article  MathSciNet  Google Scholar 

  • Ko, C.W., Lee, J., Queyranne, M.: An exact algorithm for maximum entropy sampling. Oper. Res. 43, 684–691 (1995)

    Article  MathSciNet  MATH  Google Scholar 

  • Lindsey, D.V.: On a measure of the information provided by an experiment. Ann. Math. Stat. 27, 986–1005 (1956)

    Article  Google Scholar 

  • Lindsey, B.G., Yi, G.Y., Sum, J.: Issues and strategies in the selection of composite likelihoods. Stat. Sin. 21, 71–105 (2011)

    Google Scholar 

  • Sebastiani, P., Wynn, H.P.: Maximum entropy sampling and optimal Bayesian experimental design. J. R. Stat. Soc. B 62, 145–157 (2000)

    Article  MathSciNet  MATH  Google Scholar 

  • Singer, A.: Maximum entropy formulation of the Kirkwood superposition approximation. J. Chem. Phys. 121, 3657–3666 (2004)

    Article  Google Scholar 

  • Van der Vaart, A.W.: Asymptotic Statistics. Cambridge University Press, New York (1998)

    Book  MATH  Google Scholar 

  • Varin, C., Reid, N., Firth, D.: An overview of composite likelihood methods. Stat. Sin. 21, 5–42 (2011)

    MathSciNet  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Davide Ferrari .

Editor information

Editors and Affiliations

Appendix

Appendix

Proof of Proposition 1.

Let Z=(Z 1,…,Z p ) be a p-dimensional random vector with distribution p(z). Singer (2004) shows that if Z j is independent of Z k for any jk, then \(p(\mathbf{z}) = \tilde{p}(\mathbf{z})^{(p)} = \prod_{|E|<p} p_{E}(\mathbf{z}_{E})^{q_{E}}\), where E is a set in the power set of indexes , |E| is the cardinality of the set E, q E=(−1)p+1−|E|, and p E denotes the distribution of Z E Z. Without loss of generality, we start from θ 1, θ 2 and θ 3 and write

$$\log p(\theta_1, \theta_2, \theta_3|y, \xi) = \log\tilde{p}(\varTheta|y,\xi)^{(2)} + \log p(\theta_1| \theta_2, \theta_3, y, \xi) - \log p( \theta_1| \theta_2, y, \xi). $$

Recursively applying the formula by Singer (2004) for 3≤kp, gives

(8)

By summing over all such decompositions and taking the expectation with respect to Θ|Y,ξ, we obtain

which implies \(L(\xi) = \operatorname {E}_{Y} H(\varTheta|Y, \xi) = L^{(p)}(\xi) + S_{p}(\varTheta|Y,\xi) \). Finally, by our sparsity assumption, the last summand converges to zero as p→∞.

Rights and permissions

Reprints and permissions

Copyright information

© 2013 Springer International Publishing Switzerland

About this paper

Cite this paper

Ferrari, D., Borrotti, M. (2013). Maximum Entropy Design in High Dimensions by Composite Likelihood Modelling. In: Ucinski, D., Atkinson, A., Patan, M. (eds) mODa 10 – Advances in Model-Oriented Design and Analysis. Contributions to Statistics. Springer, Heidelberg. https://doi.org/10.1007/978-3-319-00218-7_9

Download citation

Publish with us

Policies and ethics