학술논문

Probability theory.
Document Type
Book Review
Author
Jaynes, E. T. AMS Author Profile
Source
Subject
60 Probability theory and stochastic processes
  60-02 Research exposition
Language
English
Abstract
The theme of this book is summarized in a footnote on p.\ xix: ``our topic is the optimal processing of incomplete information''. And again, ``Our theme is simply: probability theory as extended logic'' [p.\ xxii]. Roughly, and perhaps slightly inaccurately put, this work is a development of probability and statistics from this point of view. \par While frequentists might easily dismiss this as another Bayesian text, statisticians of the Bayesian persuasion would have greater difficulty in seeing it as supporting more classical (Fisher, Neyman-Pearson) methods. Jaynes himself, however, has no hesitation, claiming that his method, based on the principle of maximum entropy, encompasses not only Bayesian and frequency calculations but also calculations in neither group. \par The book is in two parts: I. Principles and elementary applications, and II. Advanced applications. There are a number of challenging exercises, two indexes, a list of references and a bibliography. \par In Chapter 1, ``Plausible reasoning'', Jaynes introduces the keystone of his theory: a robot that will ``carry out useful plausible reasoning, following clearly defined principles expressing an idealized common sense'' [p.\ 8]. Three basic desiderata are given, which show not only the inner workings of the robot's brain, but also how the robot relates (or, more accurately, should relate) to the external world. \par In Chapter 2 the usual fundamental rules for probabilities are derived from the aforementioned desiderata. The method follows that of R. T. Cox\ [{\it The algebra of probable inference}, The Johns Hopkins Press, Baltimore, Md, 1961; MR0130703 (24 \#A563)]. \par ``Elementary sampling theory'' is the subject of Chapter 3, while Chapter 4 is concerned with elementary hypothesis testing. The reader used to a more conventional approach may be surprised to find at the beginning of this chapter a discussion of likelihoods, prior and posterior probabilities. \par In Chapter 5, ``Queer uses for probability theory'', matters as diverse as extra-sensory perception, the discovery of Neptune, horse racing and weather forecasting are considered. \par Chapter 6 is concerned with elementary parameter estimation. In $\S6.20$ Jaynes discusses the ``taxicab problem'' and mentions his inability to find it in the orthodox literature. It is in fact discussed in [H. Jeffreys, {\it Theory of probability}, Third edition, Clarendon Press, Oxford, 1961; MR0187257 (32 \#4710)] (where it is concerned with tramcars) and in [C. P. Robert, {\it The Bayesian choice}, Translated and revised from the French original by the author, Springer, New York, 1994; MR1313727 (97d:62059)]. Notice, by the way, that while Jeffreys attributes the problem to Maxwell Herman Alexander Newman, a gremlin---dare we say Maxwell's demon?---causes Robert to attribute it to [Jerzy?] Neyman! \par Chapter 7 is concerned with the normal distribution, and Chapter 8 deals with sufficiency, ancillarity ``and all that''. \par ``Repetitive experiments: probability and frequency'' is the title of the ninth chapter, one in which Jaynes examines what predictions his robot can make with very limited information. \par In Chapter 10 Jaynes considers the fundamental difficulty associated with the notion of ``random'' experiments, particularly in relation to physics. The entropy principle is introduced in Chapter 11, the first in the advanced applications section. Chapter 12 is concerned with ignorance priors and transformation groups. \par Bertrand's paradox is discussed here, the recognition of an element of rotational symmetry enabling Jaynes to arrive at a unique answer. Jaynes had considered this paradox in 1973 [Found. Phys. {\bf 3} (1973), 477--492; MR0426227 (54 \#14173)], and it is perhaps a pity that he does not respond here to later relevant work by others---e.g.\ K. S. Friedman [Found.\ Phys. {\bf 5} (1975), no. 1, 89--91] and A. Nathan\ [Philos. Sci. {\bf 51} (1984), no.~4, 677--684 781314 ]. \par Jaynes devotes his thirteenth chapter, ``Decision theory, historical background'', to showing that the procedures he has so far introduced are optimal in terms of some clearly defined criterion. In Chapter 14 the decision theory of the previous chapter is applied to the detection of signals in noise and the detection of systematic drift in machine characteristics. \par The word ``paradox'' is ambiguous, and Jaynes advisedly gives a precise definition, perhaps different from that in common acceptance, of his usage of the term at the start of his fifteenth chapter, one devoted to some probabilistic paradoxes. \par The next chapter is devoted to the historical background of orthodox statistics. Jaynes being a physicist, we can perhaps see his view on scientific inference, similar to that of Daniel Bernoulli, Laplace and other mathematical physicists, as tainted, or at least tinctured, by professional interest. Jaynes briefly discusses the methods introduced by the Fisher and Neyman-Pearson schools, and mentions the attacks and defences of men like M. G. Kendall, Cramér, Feller, von Mises, ``and even the putative Bayesian L. J. Savage'' [p.\ 493], etc.---the certainty expressed by these savants in their discussions Jaynes finds almost inexplicable ``since they were all quite competent mathematically'' [p.\ 493]. O mysterious condescension! \par In Chapter 17, the ``Principles and pathology of orthodox statistics'', Jaynes investigates the consequences of not using Bayesian methods in some very simple problems. Cox's theorems, as developed by Jaynes, facilitate the examination of how, and under what circumstances, orthodox and Bayesian results differ. \par In Chapter 18, ``The $A_p$ distribution and rule of succession'', Jaynes investigates whether some machinery can be built into the robot that will enable it to store general conclusions. This is effected by the introduction of a new proposition $A_p$, interpreted as ``regardless of anything else you may have been told, the probability of $A$ is $p$''. \par Jaynes returns to the matter of physical measurement in Chapter 19, paying especial attention to the problem of the great inequality of Jupiter and Saturn. Chapter 20 is concerned with model comparison, and ``Outliers and robustness'' is the subject of Chapter 21. The final chapter presents an introduction to communication theory. \par That both the references and bibliography are annotated is most useful, Jaynes showing in some of his remarks that he is indeed fortiter in both {\it re} and {\it modo}. \par The anecdote about ``Given $2+2=5$: prove that I am the Pope'' attributed in Appendix B to G. H. Hardy and J. E. McTaggart is ascribed in [G. Birkhoff, {\it Lattice theory}, Corrected reprint of the 1967 third edition, Amer. Math. Soc., Providence, R.I., 1979; MR0598630 (82a:06001)] to Bertrand Russell. Further, note that Borel (1924) is in fact translated in the first edition (1964) of Kyburg \& Smokler, and is not in the second. \par Jaynes notes that part of the problem between Jeffreys (a geophysicist) and Fisher (a geneticist) rose from their consideration of very different problems. One may well ask oneself whether there is {\it one} probabilistic/statistical methodology that is appropriate for all scientific inference, as this book in fact suggests. \par This is a work written by a scientist for scientists. As such it is to be welcomed. The reader will certainly find things with which he disagrees: but he will also find much that will cause him to think deeply not only on his usual practice but also on statistics and probability in general. {\it Probability theory: the logic of science} is, for both statisticians and scientists, more than just ``recommended reading'': it should be prescribed.

Online Access