Skip to main content
Log in

Information combination and transformation: a paradox and its resolution

  • Research
  • Published:
Accreditation and Quality Assurance Aims and scope Submit manuscript

Abstract

In his recent paper (https://doi.org/10.1016/j.measen.2022.100416), Willink showed two contradictory theoretical solutions to the problem of combining and transforming two sets of information about the same quantity. Each set of information is represented by a probability density function (PDF), and the transformation function is nonlinear. We refer to Willink’s contradictory result as “Willink paradox”. Two operations: (a) information combination and (b) information transformation are both mathematically valid according to probability theory. Therefore, the two contradictory solutions are both theoretically correct. In practice such as measurement uncertainty analysis, however, we cannot use both solutions; we must choose one or the other. We propose an entropy metric that can be used to provide a practical solution to the Willink paradox. Three examples are presented to illustrate the Willink paradox and the proposed entropy metric.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3

Similar content being viewed by others

References

  1. Huang H (2020) Two simple and practical methods for combining prior information with current measurement in uncertainty analysis. Cal Lab Int J Metrol 27(3):22–32

    Google Scholar 

  2. Shannon CE (1948) A mathematical theory of communications. Bell Syst Tech J 27(379–423):623–656

    Article  Google Scholar 

  3. Willink R (2022) On revision of the guide to the expression of uncertainty in measurement: proofs of fundamental errors in Bayesian approaches. Meas: Sens. https://doi.org/10.1016/j.measen.2022.100416

    Article  Google Scholar 

  4. Zhou V (2019) A simple explanation of information gain and entropy (updated September 2022) https://victorzhou.com/blog/information-gain/

Download references

Acknowledgements

The author would like to thank Dr. Robin Willink and an anonymous reviewer for their valuable comments that helped improve the quality of this paper.

Author information

Authors and Affiliations

Authors

Contributions

HH wrote the manuscript text, prepared all figures, and reviewed the manuscript.

Corresponding author

Correspondence to Hening Huang.

Ethics declarations

Conflict of interest

The authors declare no conflict of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Huang, H. Information combination and transformation: a paradox and its resolution. Accred Qual Assur (2024). https://doi.org/10.1007/s00769-024-01578-x

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s00769-024-01578-x

Keywords

Navigation