Check Out Science Podcasts at Blog Talk Radio with Virtually Speaking Science on BlogTalkRadio
Wednesday, May 06, 2015
Alan Boyle & Matt Strassler • Large Hadron Collider 05/06 by Virtually Speaking Science | Science Podcasts
Alan Boyle & Matt Strassler • Large Hadron Collider 05/06 by Virtually Speaking Science | Science Podcasts
Thursday, November 20, 2014
Information about the event was blogged by Professor Matt Strassler at, At the Naturalness 2014 Conference. See also his explanation on Naturalness and the Standard Model
Information about the event itself.
Information about the event itself.
The discovery of a Higgs boson, with a mass around 125 GeV, at the LHC is a great victory for the Standard Model (SM). With its minimal scalar sector of electroweak symmetry breaking, the SM at short distances, well below the proton radius, is a complete weakly coupled theory. Even though the SM cannot explain several experimental observations such as neutrino masses, the baryon asymmetry of the universe and the origin of dark matter, one cannot deduce an energy scale at which the SM would be forced to be extended (with the exceptions of the Planck scale and the Landau pole of the hypercharge force). See: Naturalness 2014
In physics, naturalness is the property that the free parameters or physical constants appearing in a physical theory should take relative values "of order 1". That is, a natural theory would have parameters with values like 2.34 rather than 234,000 or 0.000234. This is in contrast to current theory like the standard model, where there are a number of parameters that vary by many orders of magnitude, and require extensive "fine-tuning" of those values in order for the theory to predict a universe like the one we live in.
The requirement that satisfactory theories should be "natural" in this sense is a current of thought initiated around the 1960s in particle physics. It is an aesthetic criterion, not a physical one, that arises from the seeming non-naturalness of the standard model and the broader topics of the hierarchy problem, fine-tuning, and the anthropic principle.
It is not always compatible with Occam's razor, since many instances of "natural" theories have more parameters than "fine-tuned" theories such as the Standard Model.
Now that naturalism has become an accepted component of philosophy, there has recently been interest in Kuhn's work in the light of developments in the relevant sciences, many of which provide corroboration for Kuhn's claim that science is driven by relations of perceived similarity and analogy. It may yet be that a characteristically Kuhnian thesis will play a prominent part in our understanding of science http://plato.stanford.edu/entries/thomas-kuhn/
A non-technical discussion of the naturalness criterion and its implications for new physics searches at the LHC. To be published in the book "LHC Perspectives", edited by G. Kane and A. Pierce. See: Naturally Speaking: The Naturalness Criterion and Physics at the LHC(PDF)
Tuesday, March 12, 2013
Sean Carroll, Matt Strassler & Alan Boyle 02/06 by Virtually Speaking Science | Blog Talk Radio
Sean Carroll & Matt Strassler talk particle physics, time and space with Host Alan Boyle.
Tuesday, September 04, 2012
Here are a series of written Blog entries by Matt Strassler from his Blog, Of Particular Significance.
- Ball on a Spring (Classical)
- Ball on a Spring (Quantum)
- Waves (Classical Form)
- Waves (Classical Equation of Motion)
- Waves (Quantum)
- Particles are Quanta
- How fields and particles interact with each other
- How the Higgs Field Works
Given a preceding map by Proffessor Strassler according to what has been gain in finality views requires this updating in order to proceed correctly in the views shared currently in science. So that lineage of thought is important to me.
At the same time one cannot be held back from looking further and seeing where theoretical views have been taken beyond the constraints applied to the science mind.:)
Pythagoras could be called the first known string theorist. Pythagoras, an excellent lyre player, figured out the first known string physics -- the harmonic relationship. Pythagoras realized that vibrating Lyre strings of equal tensions but different lengths would produce harmonious notesratio of the lengths of the two strings were a whole number. (i.e. middle C and high C) if the......
Pythagoras discovered this by looking and listening. Today that information is more precisely encoded into mathematics, namely the wave equation for a string with a tension T and a mass per unit length m. If the string is described in coordinates as in the drawing below, where x is the distance along the string and y is the height of the string, as the string oscillates in time t,
See: Official String Theory Web Site
Friday, January 13, 2012
One string theorist even went so far to conclude that a string theory calculation of Kaluza-Klein modes was the "smoking gun" that proved our theory was the same as the string theory that string theorists had already been studying.Warped Passages: Unraveling The Mysteries of the Universes Hidden Dimensions by Lisa Randall Pg 436, Para 4
Putting this together with what is real in our reality is of importance as well. While I may have my own metaphysical development and model building characteristics it was important that I learn the scientific one so that I could see where I may have been wrong in my own development scenario. Wrong in my own intuitions.
Meanwhile I’m continuing to develop the Extra Dimensions series of articles, and I’ve now followed up my examples of extra dimensions with a next installment, a first discussion of what scientists would look for in trying to identify that our world actually has one or more extra dimensions . The new article describes one of the key clues that would indicate their presence. But this is far from the end of the story: I owe you more articles, explaining why extra dimensions would generate this clue, outlining how we try to search for this clue experimentally, and mentioning other possible clues that might arise. All in due course…The Smoking Gun for Extra Dimensions by Theoretical Physicist Matt Strassler
Some may of not been forced to question them-self with what it is that we have to ask of ourselves, as we delve into the world of the sciences and philosophies. To ask ourselves whether we had always been dealing with the truth of our getting to the heart of things.
A professor may have asked what it is exactly what I wanted out of all of this, and to him I have to relay a dream that has manifested because of his question.
In the dream I have been provided a forum for discussing my ideas.....but when it came to the time for speaking, my preparations, I felt lost as to where to begin. So it seems I have come to this point in time, as to "shit or get off the pot" as to what it is I wish to share of importance?
Giving these subjects the numbers of years since 2001, one would have thought had served my own internship, but alas I remain ever the student with no classification. Yet it is the developing of the concepts with what is real in the push to experiment as to find what the real world examples are showing as attributes in the experimental processes as they unfold.
In this example I’m going to map speed to the pitch of the note, length/postion to the duration of the note and number of turns/legs/puffs to the loudness of the note.See: How to make sound out of anything.
Who of us has the foresight to see where the process of the experiment had been developed to share an idea about what it was that we wanted to discover of nature? To see in the mind of the developers as to why the equipment has been superimposed from the schematics of theories to be tested as to discover what we may found in our model building.
Does all this prepare you to looking at the universe different?
The Lagrange Points
In the above contour plot we see that L4 and L5 correspond to hilltops and L1, L2 and L3 correspond to saddles (i.e. points where the potential is curving up in one direction and down in the other). This suggests that satellites placed at the Lagrange points will have a tendency to wander off (try sitting a marble on top of a watermelon or on top of a real saddle and you get the idea). A detailed analysis (PDF link) confirms our expectations for L1, L2 and L3, but not for L4 and L5. When a satellite parked at L4 or L5 starts to roll off the hill it picks up speed. At this point the Coriolis force comes into play - the same force that causes hurricanes to spin up on the earth - and sends the satellite into a stable orbit around the Lagrange point. See: Space Travel and Propulsion Methods
I have to say who has not been touched as if we put on a pair of rose colored glasses to see the Lagrangian world as if the gravitons populated locations of influence. As if they were descriptive as overlapping nodes of sound as to support some acoustical idea about levitation? Satellites that travel through space or held in position as our space station is.
Like different musical instruments, different types of stars produce different types of sound waves. Small stars produce a sound with a higher pitch than bigger stars, just like the 'piccolo' produces a higher sound than the cello
Thus it is as ones can see differently that I look upon the world as to discover what things we may not know of our own selves that we had missed in understanding our own physical evolution, that it is more then the matter with which we use and are made up of?
This recording was produced by converting into audible sounds some of the radar echoes received by Huygens during the last few kilometres of its descent onto Titan. As the probe approaches the ground, both the pitch and intensity increase. Scientists will use intensity of the echoes to speculate about the nature of the surface. Radar echos from Titan's surface
Monday, December 19, 2011
practically nobody took very seriously the CDF claim.......Tommaso: I will claim based on the above that according to Prof. D'Agostini, Prof. Matt Strassler is "practically nobody", since he is not convinced.
An acoustical difference of opinion with regard too, Nobody?
possible phrases spoken?
Hmmmmm.......One is a shop keeper and one is a customer?
Bayesian probability is one of the different interpretations of the concept of probability and belongs to the category of evidential probabilities. The Bayesian interpretation of probability can be seen as an extension of logic that enables reasoning with propositions whose truth or falsity is uncertain. To evaluate the probability of a hypothesis, the Bayesian probabilist specifies some prior probability, which is then updated in the light of new, relevant data.
The Bayesian interpretation provides a standard set of procedures and formulae to perform this calculation. Bayesian probability interprets the concept of probability as " a probability p is an abstract concept, a quantity that we assign theoretically, for the purpose of representing a state of knowledge, or that we calculate from previously assigned probabilities," in contrast to interpreting it as a frequency or a "propensity" of some phenomenon.
The term "Bayesian" refers to the 18th century mathematician and theologian Thomas Bayes (1702–1761), who provided the first mathematical treatment of a non-trivial problem of Bayesian inference. Nevertheless, it was the French mathematician Pierre-Simon Laplace (1749–1827) who pioneered and popularised what is now called Bayesian probability.
Broadly speaking, there are two views on Bayesian probability that interpret the probability concept in different ways. According to the objectivist view, the rules of Bayesian statistics can be justified by requirements of rationality and consistency and interpreted as an extension of logic. According to the subjectivist view, probability measures a "personal belief". Many modern machine learning methods are based on objectivist Bayesian principles. In the Bayesian view, a probability is assigned to a hypothesis, whereas under the frequentist view, a hypothesis is typically tested without being assigned a probability.
Bayesian methodologyIn general, Bayesian methods are characterized by the following concepts and procedures:
- The use of hierarchical models and marginalization over the values of nuisance parameters. In most cases, the computation is intractable, but good approximations can be obtained using Markov chain Monte Carlo methods.
- The sequential use of the Bayes' formula: when more data become available after calculating a posterior distribution, the posterior becomes the next prior.
- In frequentist statistics, a hypothesis is a proposition (which must be either true or false), so that the (frequentist) probability of a frequentist hypothesis is either one or zero. In Bayesian statistics, a probability can be assigned to a hypothesis.
Objective and subjective Bayesian probabilitiesBroadly speaking, there are two views on Bayesian probability that interpret the 'probability' concept in different ways. For objectivists, probability objectively measures the plausibility of propositions, i.e. the probability of a proposition corresponds to a reasonable belief everyone (even a "robot") sharing the same knowledge should share in accordance with the rules of Bayesian statistics, which can be justified by requirements of rationality and consistency. Requirements of rationality and consistency are also important for subjectivists, for which the probability corresponds to a 'personal belief'. For subjectivists however, rationality and consistency constrain the probabilities a subject may have, but allow for substantial variation within those constraints. The objective and subjective variants of Bayesian probability differ mainly in their interpretation and construction of the prior probability.
HistoryThe term Bayesian refers to Thomas Bayes (1702–1761), who proved a special case of what is now called Bayes' theorem in a paper titled "An Essay towards solving a Problem in the Doctrine of Chances". In that special case, the prior and posterior distributions were Beta distributions and the data came from Bernoulli trials. It was Pierre-Simon Laplace (1749–1827) who introduced a general version of the theorem and used it to approach problems in celestial mechanics, medical statistics, reliability, and jurisprudence. Early Bayesian inference, which used uniform priors following Laplace's principle of insufficient reason, was called "inverse probability" (because it infers backwards from observations to parameters, or from effects to causes). After the 1920s, "inverse probability" was largely supplanted by a collection of methods that came to be called frequentist statistics.
In the 20th century, the ideas of Laplace were further developed in two different directions, giving rise to objective and subjective currents in Bayesian practice. In the objectivist stream, the statistical analysis depends on only the model assumed and the data analysed. No subjective decisions need to be involved. In contrast, "subjectivist" statisticians deny the possibility of fully objective analysis for the general case.
In the 1980s, there was a dramatic growth in research and applications of Bayesian methods, mostly attributed to the discovery of Markov chain Monte Carlo methods, which removed many of the computational problems, and an increasing interest in nonstandard, complex applications. Despite the growth of Bayesian research, most undergraduate teaching is still based on frequentist statistics. Nonetheless, Bayesian methods are widely accepted and used, such as in the fields of machine learning and talent analytics.
Justification of Bayesian probabilitiesThe use of Bayesian probabilities as the basis of Bayesian inference has been supported by several arguments, such as the Cox axioms, the Dutch book argument, arguments based on decision theory and de Finetti's theorem.
Axiomatic approachRichard T. Cox showed that Bayesian updating follows from several axioms, including two functional equations and a controversial hypothesis of differentiability. It is known that Cox's 1961 development (mainly copied by Jaynes) is non-rigorous, and in fact a counterexample has been found by Halpern. The assumption of differentiability or even continuity is questionable since the Boolean algebra of statements may only be finite. Other axiomatizations have been suggested by various authors to make the theory more rigorous.
Dutch book approachThe Dutch book argument was proposed by de Finetti, and is based on betting. A Dutch book is made when a clever gambler places a set of bets that guarantee a profit, no matter what the outcome is of the bets. If a bookmaker follows the rules of the Bayesian calculus in the construction of his odds, a Dutch book cannot be made.
However, Ian Hacking noted that traditional Dutch book arguments did not specify Bayesian updating: they left open the possibility that non-Bayesian updating rules could avoid Dutch books. For example, Hacking writes "And neither the Dutch book argument, nor any other in the personalist arsenal of proofs of the probability axioms, entails the dynamic assumption. Not one entails Bayesianism. So the personalist requires the dynamic assumption to be Bayesian. It is true that in consistency a personalist could abandon the Bayesian model of learning from experience. Salt could lose its savour."
In fact, there are non-Bayesian updating rules that also avoid Dutch books (as discussed in the literature on "probability kinematics" following the publication of Richard C. Jeffrey's rule). The additional hypotheses sufficient to (uniquely) specify Bayesian updating are substantial, complicated, and unsatisfactory.
Decision theory approachA decision-theoretic justification of the use of Bayesian inference (and hence of Bayesian probabilities) was given by Abraham Wald, who proved that every admissible statistical procedure is either a Bayesian procedure or a limit of Bayesian procedures. Conversely, every Bayesian procedure is admissible.
Personal probabilities and objective methods for constructing priorsFollowing the work on expected utility theory of Ramsey and von Neumann, decision-theorists have accounted for rational behavior using a probability distribution for the agent. Johann Pfanzagl completed the Theory of Games and Economic Behavior by providing an axiomatization of subjective probability and utility, a task left uncompleted by von Neumann and Oskar Morgenstern: their original theory supposed that all the agents had the same probability distribution, as a convenience. Pfanzagl's axiomatization was endorsed by Oskar Morgenstern: "Von Neumann and I have anticipated" the question whether probabilities "might, perhaps more typically, be subjective and have stated specifically that in the latter case axioms could be found from which could derive the desired numerical utility together with a number for the probabilities (cf. p. 19 of The Theory of Games and Economic Behavior). We did not carry this out; it was demonstrated by Pfanzagl ... with all the necessary rigor".
Ramsey and Savage noted that the individual agent's probability distribution could be objectively studied in experiments. The role of judgment and disagreement in science has been recognized since Aristotle and even more clearly with Francis Bacon. The objectivity of science lies not in the psychology of individual scientists, but in the process of science and especially in statistical methods, as noted by C. S. Peirce. Recall that the objective methods for falsifying propositions about personal probabilities have been used for a half century, as noted previously. Procedures for testing hypotheses about probabilities (using finite samples) are due to Ramsey (1931) and de Finetti (1931, 1937, 1964, 1970). Both Bruno de Finetti and Frank P. Ramsey acknowledge their debts to pragmatic philosophy, particularly (for Ramsey) to Charles S. Peirce.
The "Ramsey test" for evaluating probability distributions is implementable in theory, and has kept experimental psychologists occupied for a half century. This work demonstrates that Bayesian-probability propositions can be falsified, and so meet an empirical criterion of Charles S. Peirce, whose work inspired Ramsey. (This falsifiability-criterion was popularized by Karl Popper.)
Modern work on the experimental evaluation of personal probabilities uses the randomization, blinding, and Boolean-decision procedures of the Peirce-Jastrow experiment. Since individuals act according to different probability judgements, these agents' probabilities are "personal" (but amenable to objective study).
Personal probabilities are problematic for science and for some applications where decision-makers lack the knowledge or time to specify an informed probability-distribution (on which they are prepared to act). To meet the needs of science and of human limitations, Bayesian statisticians have developed "objective" methods for specifying prior probabilities.
Indeed, some Bayesians have argued the prior state of knowledge defines the (unique) prior probability-distribution for "regular" statistical problems; cf. well-posed problems. Finding the right method for constructing such "objective" priors (for appropriate classes of regular problems) has been the quest of statistical theorists from Laplace to John Maynard Keynes, Harold Jeffreys, and Edwin Thompson Jaynes: These theorists and their successors have suggested several methods for constructing "objective" priors:
Each of these methods contributes useful priors for "regular" one-parameter problems, and each prior can handle some challenging statistical models (with "irregularity" or several parameters). Each of these methods has been useful in Bayesian practice. Indeed, methods for constructing "objective" (alternatively, "default" or "ignorance") priors have been developed by avowed subjective (or "personal") Bayesians like James Berger (Duke University) and José-Miguel Bernardo (Universitat de València), simply because such priors are needed for Bayesian practice, particularly in science. The quest for "the universal method for constructing priors" continues to attract statistical theorists.
Thus, the Bayesian statistican needs either to use informed priors (using relevant expertise or previous data) or to choose among the competing methods for constructing "objective" priors.
- Bertrand's paradox: a paradox in classical probability, solved by E.T. Jaynes in the context of Bayesian probability
- De Finetti's game – a procedure for evaluating someone's subjective probability
- An Essay towards solving a Problem in the Doctrine of Chances
- ^ Paulos, John Allen. The Mathematics of Changing Your Mind, New York Times (US). August 5, 2011; retrieved 2011-08-06
- ^ a b c Jaynes, E.T. "Bayesian Methods: General Background." In Maximum-Entropy and Bayesian Methods in Applied Statistics, by J. H. Justice (ed.). Cambridge: Cambridge Univ. Press, 1986
- ^ Stigler, Stephen M. (1986) The history of statistics. Harvard University press. pg 131.
- ^ Stigler, Stephen M. (1986) The history of statistics., Harvard University press. pg 97-98, pg 131.
- ^ a b c Cox, Richard T. Algebra of Probable Inference, The Johns Hopkins University Press, 2001
- ^ a b de Finetti, B. (1974) Theory of probability (2 vols.), J. Wiley & Sons, Inc., New York
- ^ a b Bishop, C.M. Pattern Recognition and Machine Learning. Springer, 2007
- ^ McGrayne, Sharon Bertsch. (2011). The Theory That Would Not Die, p. 10. at Google Books
- ^ Stigler, Stephen M. (1986) The history of statistics. Harvard University press. Chapter 3.
- ^ a b Fienberg, Stephen. E. (2006) When did Bayesian Inference become "Bayesian"? Bayesian Analysis, 1 (1), 1–40. See page 5.
- ^ Bernardo, J.M. (2005), Reference analysis, Handbook of statistics, 25, 17–90
- ^ Wolpert, R.L. (2004) A conversation with James O. Berger, Statistical science, 9, 205–218
- ^ Bernardo, José M. (2006) A Bayesian mathematical statistics primer. ICOTS-7
- ^ Halpern, J. A counterexample to theorems of Cox and Fine, Journal of Artificial Intelligence Research, 10: 67-85.
- ^ a b Dupré, Maurice J., Tipler, Frank T. New Axioms For Bayesian Probability, Bayesian Analysis (2009), Number 3, pp. 599-606
- ^ Hacking (1967, Section 3, page 316), Hacking (1988, page 124)
- ^ van Frassen, B. (1989) Laws and Symmetry, Oxford University Press. ISBN 0198248601
- ^ Wald, Abraham. Statistical Decision Functions. Wiley 1950.
- ^ Bernardo, José M., Smith, Adrian F.M. Bayesian Theory. John Wiley 1994. ISBN 0-471-92416-4.
- ^ Pfanzagl (1967, 1968)
- ^ Morgenstern (1976, page 65)
- ^ Davidson et al. (1957)
- ^ "Karl Popper" in Stanford Encyclopedia of Philosophy
- ^ Popper, Karl. (2002) The Logic of Scientific Discovery 2nd Edition, Routledge ISBN 0415278430 (Reprint of 1959 translation of 1935 original) Page 57.
- ^ Pierce & Jastrow (1885)
- ^ a b Bernardo, J. M. (2005). Reference Analysis. Handbook of Statistics 25 (D. K. Dey and C. R. Rao eds). Amsterdam: Elsevier, 17-90
- Berger, James O. (1985). Statistical Decision Theory and Bayesian Analysis. Springer Series in Statistics (Second ed.). Springer-Verlag. ISBN 0-387-96098-8.
- Bernardo, José M.; Smith, Adrian F. M. (1994). Bayesian Theory. Wiley. ISBN 047149464x.
- Bickel, Peter J.; Doksum, Kjell A. (2001). Mathematical statistics, Volume 1: Basic and selected topics (Second (updated printing 2007) of the Holden-Day 1976 ed.). Pearson Prentice–Hall. ISBN 013850363X. MR443141.
- Davidson, Donald; Suppes, Patrick; Siegel, Sidney (1957). Decision-Making: An Experimental Approach. Stanford University Press.
- de Finetti, Bruno. "Probabilism: A Critical Essay on the Theory of Probability and on the Value of Science," (translation of 1931 article) in Erkenntnis, volume 31, September 1989.
- de Finetti, Bruno (1937) "La Prévision: ses lois logiques, ses sources subjectives," Annales de l'Institut Henri Poincaré,
- de Finetti, Bruno. "Foresight: its Logical Laws, Its Subjective Sources," (translation of the 1937 article in French) in H. E. Kyburg and H. E. Smokler (eds), Studies in Subjective Probability, New York: Wiley, 1964.
- de Finetti, Bruno (1974–5). Theory of Probability. A Critical Introductory Treatment, (translation by A.Machi and AFM Smith of 1970 book) 2 volumes. Wiley ISBN 0471201413, ISBN 0471201421
- DeGroot, Morris (2004) Optimal Statistical Decisions. Wiley Classics Library. (Originally published 1970.) ISBN 0-471-68029-X.
- Hacking, Ian (December 1967). "Slightly More Realistic Personal Probability". Philosophy of Science 34 (4): 311–325. doi:10.1086/288169. JSTOR 186120. Partly reprinted in: Gärdenfors, Peter and Sahlin, Nils-Eric. (1988) Decision, Probability, and Utility: Selected Readings. 1988. Cambridge University Press. ISBN 0521336589
- Hajek, A. and Hartmann, S. (2010): "Bayesian Epistemology", in: Dancy, J., Sosa, E., Steup, M. (Eds.) (2001) A Companion to Epistemology, Wiley. ISBN 1405139005 Preprint
- Hald, Anders (1998). A History of Mathematical Statistics from 1750 to 1930. New York: Wiley. ISBN 0471179124.
- Hartmann, S. and Sprenger, J. (2011) "Bayesian Epistemology", in: Bernecker, S. and Pritchard, D. (Eds.) (2011) Routledge Companion to Epistemology. Routledge. ISBN 10415962196 (Preprint)
- Howson, C.; Urbach, P. (2005). Scientific Reasoning: the Bayesian Approach (3rd ed.). Open Court Publishing Company. ISBN 978-0812695786.
- Jaynes E.T. (2003) Probability Theory: The Logic of Science, CUP. ISBN 9780521592710 (Link to Fragmentary Edition of March 1996).
- McGrayne, Sharon Bertsch. (2011). The Theory That Would Not Die: How Bayes' Rule Cracked The Enigma Code, Hunted Down Russian Submarines, & Emerged Triumphant from Two Centuries of Controversy. New Haven: Yale University Press. 13-ISBN 9780300169690/10-ISBN 0300169698; OCLC 670481486
- Morgenstern, Oskar (1978). "Some Reflections on Utility". In Andrew Schotter. Selected Economic Writings of Oskar Morgenstern. New York University Press. pp. 65–70. ISBN 9780814777718.
- Pierce, C.S. and Jastrow J. (1885). "On Small Differences in Sensation". Memoirs of the National Academy of Sciences 3: 73–83.
- Pfanzagl, J (1967). "Subjective Probability Derived from the Morgenstern-von Neumann Utility Theory". In Martin Shubik. Essays in Mathematical Economics In Honor of Oskar Morgenstern. Princeton University Press. pp. 237–251.
- Pfanzagl, J. in cooperation with V. Baumann and H. Huber (1968). "Events, Utility and Subjective Probability". Theory of Measurement. Wiley. pp. 195–220.
- Ramsey, Frank Plumpton (1931) "Truth and Probability" (PDF), Chapter VII in The Foundations of Mathematics and other Logical Essays, Reprinted 2001, Routledge. ISBN 0415225469,
- Stigler, Stephen M. (1990). The History of Statistics: The Measurement of Uncertainty before 1900. Belknap Press/Harvard University Press. ISBN 0-674-40341-X.
- Stigler, Stephen M. (1999) Statistics on the Table: The History of Statistical Concepts and Methods. Harvard University Press. ISBN 0-674-83601-4
- On-line textbook: Information Theory, Inference, and Learning Algorithms, by David MacKay, has many chapters on Bayesian methods, including introductory examples; arguments in favour of Bayesian methods (in the style of Edwin Jaynes); state-of-the-art Monte Carlo methods, message-passing methods, and variational methods; and examples illustrating the intimate connections between Bayesian inference and data compression.
- An Intuitive Explanation of Bayesian Reasoning A very gentle introduction by Eliezer Yudkowsky
- An on-line introductory tutorial to Bayesian probability from Queen Mary University of London
- James Franklin The Science of Conjecture: Evidence and Probability Before Pascal, history from a Bayesian point of view.
- Bayes’ Rule: A Tutorial Introduction by JV Stone.