Abstract
In the last couple of decades, a logician or a philosopher has run a risk whenever he has put the term ‘information’ into the title of one of his papers. In these days, the term ‘information’ often creates an expectation that the paper has something to do with that impressive body of results in communication theory which was first known as theory of transmission of information but which now is elliptically called information theory (in the United States at least).1 For the purposes of this paper, I shall speak of it as statistical information theory. I want to begin by making it clear that I have little to contribute to this statistical information theory as it is usually developed.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Bibliography
Y. Bar-Hillel, Language and Information, Addison-Wesley and The Jerusalem Academic Press, Reading, Mass., and Jerusalem 1964.
Y. Bar-Hillel and R. Carnap, ‘Semantic Information’. British Journal for the Philosophy of Science 4 [ 1953 ] 147 - 157.
R. Carnap, Logical Foundations of Probability, University of Chicago Press, Chicago 1950, 2nd ed., 1963.
R. Carnap, Continuum of Inductive Methods, University of Chicago Press, Chicago 1952.
R. Carnap, ‘Probability and Content Measure’, in Mind, Matter and Method (ed. by P. K. Fcyerabcnd and G. Maxwell ), University of Minnesota Press, Minneapolis 1966, pp. 248 - 260.
R. Carnap and Y. Bar-Hillel, ‘An Outline of a Theory of Semantic Information’, Technical Report No. 247, M.I.T. Research Laboratory in Electronics 1952; reprinted in [1], Ch. 15.
C. Cherry, On Human Communication, M.I.T. Press, Cambridge, Mass., 1957.
N. Goodman, Fact. Fiction, and Forecast, University of London Press, London 1955.
C. G. Hempel, ‘Inductive Inconsistencies’, Synthese 12 (1960) 439-69; reprinted in C. G. Hempel, Aspects of Scientific Explanation and Other Essays in the Philosophy of Science, The Free Press, New York 1965, pp. 53-79, especially pp. 73-78.
C. G. Hempel, ‘Deductive- nomological vs. Statistical Explanation’, in Minnesota Studies in the Philosophy of Science, vol. III [ed. by H. Feigl and G. Maxwell], University of Minnesota Press, Minneapolis 1962, pp. 98-169, especially pp. 153-156.
R. Hilpinen, ‘On Inductive Generalization in Binary Functional Calculus’, Third International Congress of Logic, Methodology and Philosophy of Science, Abstracts of Papers [mimeographed], Amsterdam 1967.
K. J. Hintikka, ‘Towards a Theory of Inductive Generalization’, in Proceedings of the 1964 International Congress for Logic, Methodology and Philosophy of Science [ed. by Y. Bar-Hillel], North-Holland Publishing Company, Amsterdam 1965, pp. 274 - 288.
K.J. Hintikka, ‘On a Combined System of Inductive Logic’, in Stadia Logico- Mathemalica et Philosophica in honorem Rolf Nevanlinna, Acta Philosophica Fennica 18 [1965] 21-30.
K. J. Hintikka, ‘A Two-Dimensional Continuum of Inductive Methods’, in Aspects of Inductive Logic (ed. by K. J. Hintikka and P. Suppes ), North-Holland Publishing Company, Amsterdam 1966, pp. 98 - 117.
K. J. Hintikka, ‘Induction by Enumeration and Induction by Elimination’, in The Problem of Inductive Logic, Proceedings of the International Colloquium in Logic, Methodology, and Philosophy of Science, London. 1965 (ed. by I. Lakatos), vol. II, North-Holland Publishing Company, Amsterdam 1967.
K. J. Hintikka, ‘The Varieties of Information and Scientific Explanation’, in Logic, Methodology and Philosophy of Science III, Proceedings of the 1967 International Congress [ed. by B. van Rootsclaar and J. F. Staal], North-Holland Publishing Company, Amsterdam 1968, pp. 151 - 171.
K. J. Hintikka and R. Hilpincn, ‘Knowledge, Acceptance, and Inductivc Logic’, in Aspects of Inductive Logic (ed. by K.J. Hintikka and P. Suppes), North- Holland Publishing Company, Amsterdam 1966, pp. 1 -20.
K. J. Hintikka and J. Pietarinen, ‘Semantic Information and Inductive Logic’, in Aspects of Inductive Logic (ed. by K. J. Hintikka and P. Suppes ), North-Holland Publishing Company, Amsterdam 1966, pp. 81 - 97.
J. G. Kemeny, ‘A Logical Measure Function’, Journal of Symbolic Logic 18 [ 1953 ] 289 - 308.
H. E. Kyburg and H. E. Smokier (eds.), Studies in Subjective Probability, John Wiley and Sons, New York 1964.
I. Levi, ‘Decision Theory and Confirmation’, Journal of Philosophy 58 (1961) 614 - 624.
R. D. Luce and H. Raiffa, Games and Decisions, John Wiley and Sons, New York 1957, pp. 318-324, especially p. 324.
E. Nagel, The Structure of Science, Harcourt, Brace and World, Inc., New York 1961.
K. R. Popper, The Logic of Scientific Discovery, Hutchinson, London 1959. The bulk of this work consists of a translation of Logik der Forschung, Springer, Vienna 1934.
K. R. Popper, ‘Degree of Confirmation’, British Journal for the Philosophy of Science 5 [1954] 143-149. [Correction, ibid. 334.]
P. A. Schilpp [ed.]. The Philosophy of Rudolf Carnap, The Library of Living Philosophers, La Salic, Illinois, 1963.
C. E. Shannon and W. Weaver, The Mathematical Theory of Communication, The University of Illinois Press, Urbana, Illinois, 1949.
References
For this theory, see e.G. [27]. A popular survey is given in [7].
On this subject, cf. [1], Chapters 15-18; also [7], pp. 229-255.
E.G., [1], p. 242.
It is therefore natural that such expected values or estimates of information as [2] should play an important role in this theory. Its statistical character is not the only reason, however, for the prevalence of expressions similar to [2]; they have a place in the semantic theory, too.
I shall draw on the works I have published elsewhere. See [12-18],
The term land the idea] goes back all the way to Boole.
Cf. [1], p. 307.
24] passim.
An example to this effect is given by Carnap in [5]. Further examples are found in my survey [16], especially the expected content p[A] — p[h], which is as likely to vary with p[h[e] as with 1 — p[/r].
See e. g. [9-10].
See e. g. [21-22].
This is in effect surmised by Hempel in [9], p. 77.
The pioneering work in this area has been done by Rudolf Carnap. See [3-4]. Cf. also [26], especially Carnap’s own contribution, pp. 966-998.
The line of thought presented here is not Carnap’s.
This measure was for a while preferred by Carnap to other probability measures, and it appears to be an especially simple one in certain respects.
Cf. my criticism of Carnap in the early pages of [12].
This was the course followed in [12].
It was proposed in [13].
This question was asked by Hintikka and Pietarinen [18], whose line of thought we shall here follow.
The choice of an appropriate utility here is a very tricky matter. Further comments on the subject are made in [16].
Notice, however, that other kinds of results might be forthcoming if we considered other kinds of relative information instead of the added surprise value in [16].
See the last few pages of [12] and [13].
These informal arguments can easily be reconstructed on the basis of the distinctions between different kinds of information discussed in [16].
For the following, cf. [4] and [14].
See e. g. Nelson Goodman [8] or E. Nagel [23], Ch. 4.
See e. g. Nelson Goodman [8] or E. Nagel [23], Ch. 4. 28 See de Finetti’s classical paper reprinted in [20].
A preliminary report on some of the simplest cases is given in [11].
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 1970 D. Reidel Publishing Company, Dordrecht-Holland
About this chapter
Cite this chapter
Hintikka, J. (1970). On Semantic Information. In: Hintikka, J., Suppes, P. (eds) Information and Inference. Synthese Library, vol 28. Springer, Dordrecht. https://doi.org/10.1007/978-94-010-3296-4_1
Download citation
DOI: https://doi.org/10.1007/978-94-010-3296-4_1
Publisher Name: Springer, Dordrecht
Print ISBN: 978-94-010-3298-8
Online ISBN: 978-94-010-3296-4
eBook Packages: Springer Book Archive