Quantum Bayesian Interpretation of QM

In summary, the QBism model presented by Fuchs & Schack is an attempt to reformulate quantum mechanics in a way that removes some of the paradoxes and inconsistencies found in the current model. However, Fields argues that the model is flawed because it does not provide a physical distinction between observers and the systems they observe, treating all quantum systems as autonomous agents that respond to observations by updating beliefs and employ quantum mechanics as a “users’ manual” to guide behavior.
  • #1
Salman2
95
0
Any comments (pro-con) on this Quantum Bayesian interpretation of QM by Fuchs & Schack ?:

http://arxiv.org/pdf/1301.3274.pdf
 
Physics news on Phys.org
  • #2
Yes, I'd also like to know if anyone has any insights on this new model, dubbed the "QBism" model. The general idea is that the quantum wave function does NOT represent any actuality in the real physical world. It is an abstraction of the mind...and it goes from there. The arxiv article Salman2 posted is called the No nonsense version, which sounds like a good initial review. However, for those who want an even briefer survey should check the most recent SciAm issue:

http://www.scientificamerican.com/a...tum-beyesnism-fix-paradoxes-quantum-mechanics
 
  • #3
"The general case of conscious perception is the negative perception, namely, 'perceiving the stone as not grey'...
"Consciousness is the subjective form involved in feeling the contrast between the 'theory' which may be erroneous and the fact which is given."

Alfred North Whitehead, Process and Reality, ISBN 0-02-934570-7, p. 161.

"In other words, consciousness enters into the subjective forms of feelings, when those feelings are components in an integral feeling whose datum is the contrast between a nexus which is, and a proposition which in its own nature negates the decision of its truth or falsehood."

p. 261

Again, H P Stapp sees Whitehead as providing a container for QM. I agree. If he had lived a little longer, he would have been in the QM Pantheon.

CW
 
Last edited:
  • #4
Chris Fields offer a critique of the QBism model presented by Fuchs based on the nature of the wavefunction of the agent (observer):

http://arxiv.org/pdf/1108.2024v2.pdf

Here from the end of the paper is the major objection of Fields to QBism:

"QBism provides no physical distinction between observers and the systems they observe, treating all quantum systems as autonomous agents that respond to observations by updating beliefs and employ quantum mechanics as a “users’ manual” to guide behavior. However, it treats observation itself as a physical process in which an “observer” acts on a “system” with a POVM and the system” selects a POVM component as the “observer’s experience” in return. This requirement renders the assumption that systems be well-defined - i.e. have constant d-impossible to implement operationally. It similarly forces the consistent QBist to regard the environment as an effectively omniscient observer, threatening the fundamental assumption of subjective probabilities and forcing the conclusion that QBist observers cannot segment their environments into objectively separate systems."

==

Another paper by Fields, discussion of QBism starts on p. 27:

http://arxiv.org/pdf/1108.4865.pdf
 
Last edited:
  • #5
Salman2 said:
Any comments (pro-con) on this Quantum Bayesian interpretation of QM by Fuchs & Schack ?:

http://arxiv.org/pdf/1301.3274.pdf

Epistemic-Epistemic view of the wave function.
 
  • #6
Subjective Reality-

The discussion of QBism poses iepistemological, and semantic problems for the reader. The subtitle- It's All ln Your Mind- is a tautology. Any theory or interpretation of observed physical phenomena is in the mind, a product of the imagination, or logical deduction, or some other mental process. Heisenberg ( The Physical Principles of the Quantum Theory), [n discussing the uncertainty principle, cautioned that human language permits the construction of sentences that have no content since they imply no experimentally observable consequences , even though they may conjure up a mental picture. He particularly cautioned against the use of the term, " real " in relation to such statements. as is done in the article. Mr. von Burgers also described QBism as representing subjective beliefs-whose? Bertrand Russell (Human Knowlrdge, Its Scope and Limits) described " belief" as a word not easy to define. It is certainly not defined in the context of the article.
,
Heisenberg also showed that the uncertainty principle and several other results of quantum mechanical theory could be deduced without reference to a wave function, so this aspect of the new interpretations is not unique. Similarly Feynman (QED, The Strange Theory of Light and Matter) dealt with the diffraction of light through a pair of slots, by a formulation based on the actions of photons without reference to wave functions. The statement in the article that the wave function is
" only a tool" to enable mathematical calculations is puzzling- any theoretical formulation of quantum mechanics is a tool for mathematical calculations relating to the properties of physical systems.

In spite of the tendency in Mr. von Burgers' article to overplay the virtues of QBism relative to other formulations, as an additional way to contemplate quantum mechanics, it has potential value, As Feynman ( The Character of Physical Law) stated, any good theoretical physicist knows six or seven theoretical representations for exactly the same physics. One or another of these may be the most advantageous way of contemplating how to extend the theory into new domains and discover new laws. Time will tell.

Alexander
 
Last edited by a moderator:
  • #7
wayseeker said:
The discussion of QBism poses iepistemological, and semantic problems for the reader. ...

Welcome to PhysicsForums, Alexander!

Are you familiar with the PBR theorem? Although I can't say I fully understand the examples in the OP's QBism paper, it seems to flow directly opposite to PBR. One says the wave function maps directly to reality, the other says it does not.
 
  • #8
Some interesting remarks on Bayes' Theorem

http://www.stat.columbia.edu/~gelman/research/published/badbayesmain.pdf
"Bayesian inference is one of the more controversial approaches to statistics, with both the promise and limitations of being a closed system of logic. There is an extensive literature, which sometimes seems to overwhelm that of Bayesian inference itself, on the advantages and disadvantages of Bayesian approaches"---------
"Bayes' Theorem is a simple formula that relates the probabilities of two different events that are conditional upon each other"

sound familiar, no ?
(in physics, i mean)
 
Last edited:
  • #9
DrChinese said:
Are you familiar with the PBR theorem? Although I can't say I fully understand the examples in the OP's QBism paper, it seems to flow directly opposite to PBR. One says the wave function maps directly to reality, the other says it does not.

I have gone through the PBR theorem and my view is exactly the same as Matt Leifer:
http://mattleifer.info/2011/11/20/can-the-quantum-state-be-interpreted-statistically/

He divides interpretations into three types:

1. Wavefunctions are epistemic and there is some underlying ontic state. Quantum mechanics is the statistical theory of these ontic states in analogy with Liouville mechanics.
2. Wavefunctions are epistemic, but there is no deeper underlying reality.
3. Wavefunctions are ontic (there may also be additional ontic degrees of freedom, which is an important distinction but not relevant to the present discussion).

PBR says nothing about type 2 - in fact the paper specifically excludes it. What it is concerned about is theories of type 1 and 3 - it basically says type 1 is untenable - its really type 3 in disguise.

That's an interesting result but I am scratching my head why its considered that important. Most interpretations are type 2 (eg Copenhagen and the Ensemble interpretation), many others are type 3 (eg MWI and BM) and only a few type 1.

Maybe I am missing something but from what I can see its not that big a deal.

Thanks
Bill
 
Last edited:
  • #10
Regarding the Quantum Baysean interpretation its a perfectly good way of coming to grips with the probability part of QM.

Normally that is done by means of an ensemble view of probability which talks about the proportion of a very large number of similar systems and the proportion with a particular property or whatever is the probability. This is a view very commonly used in applied math. But Baysian probability theory (perhaps framework is a better word) - is just as valid. In fact there is some evidence it leads to a slicker axiomatic formulation:
http://arxiv.org/pdf/quant-ph/0210017v1.pdf

Even if it isn't as slick I prefer the ensemble view because of its greater pictorial vividness.

IMHO its not really an issue to get too concerned about. In applying probability to all sorts of areas the intuitive view most have is from my experience more than adequate without being strict about it.

Thanks
Bill
 
Last edited:
  • #11
DiracPool said:
Yes, I'd also like to know if anyone has any insights on this new model, dubbed the "QBism" model. The general idea is that the quantum wave function does NOT represent any actuality in the real physical world.

As far as I can see its simply the ensemble interpretation in another guise - where the pictorial vividness of an ensemble is replaced by beliefs about information.

Information seems to be one of the buzz things in physics these days but personally I can't see the appeal - although am willing to be convinced.

You might be interested in the following where QM is derived from all systems with the same information carrying capacity are the same (plus a few other things):
http://arxiv.org/pdf/0911.0695v1.pdf

It's my favorite foundational basis of QM these days - but I have to say it leaves some cold.

The interesting thing though is if you remove information from the axioms and say - all systems that are observationally the same are equivalent it doesn't change anything in the derivation - which sort of makes you wonder.

Thanks
Bill
 
  • #12
Well, maybe I'm just too much biased by my training as a physicist to make sense of the whole Baysian interpretation of probabilities. In my opinion this has nothing to do with quantum theory but with any kind of probabilistic statement. It is also good to distinguish some simple categories of content of a physical theory.

A physical theory, if stated in a complete way like QT, is first some mathematical "game of our minds". There is a well-defined set of axioms or postulates which give a formal set of rules which establishes how to calculate abstract things. In QT that's the "state of the system", given by a self-adjoint positive semidefinite trace-class operator on a (rigged) Hilbert space, an "algebra of observables", represented by self-adjoint operators, and a Hamiltonian among the observables that defines the dynamics of the system. That's just the formal rules of the game. It's just a mathematical universe, you can make statements (prove theorems), do calculations. I think this part is totally free of interpretational issues, because no connection to the "real world" (understood as reproducible objective observations) has been made yet.

Now comes the difficult part, namely this connection with the real world, i.e., with reproducible obejective observations in nature. In my opinion, the only consistent interpretation is the Minimal Statistical Interpretation, which is basically defined by Born's Rule, saying that for a given preparation of a system in a quantum state, represented by the Statistical Operator [itex]\hat{R}[/itex] the probability (density) to measure a complete set of compatible operators is given by
[tex]P(A_1,\ldots, A_n|\hat{R})=\langle A_1,\ldots, A_n|\hat{R}|A_1,\ldots, A_n \rangle[/tex]
where [itex]|A_1,\ldots, A_n\rangle [/itex] is a (generalized) common eigenvector normalized to 1 (a [itex]\delta[/itex] distribution) of the self-adjoint operators representing the complete set of compatible operators.

Now the interpretation is shifted to the interpretation of probabilities. QT makes no other predictions about the outcome of measurements than these probabilities, and now we have to think about the meaning of probabilities. It's clear that probability theory also is given as a axiomatic set of rules (e.g., the Kolmogorov axioms), which is unproblematic since its just a mathematical abstraction. The question now is, how to interpret probabilities in the sense of physical experiments. Physics is about the test of hypothesis about real-world experiments and thus we must make this connection between probabilities and outcomes of such real-world measurements. I don't see, how else you can define this connection than by repeating the measurement on a sufficiently large ensemble of identically and independently prepared experimental setups. The larger the ensemble the higher the statistical significance for proving or disproving the predicted probabilities for the outcome of measurements.

The Bayesian view, for me, is just a play with words, trying to give a physically meaningful interpretation of probability for a single event. In practice, however, you cannot prove anything about a probabilistic statement with only looking at a single event. If I predict a probability of 10% chance of rain tomorrow, and then the fact whether it rains or doesn't rain on the next day doesn't tell anything about the validity of my probabilistic prediction. The only thing one can say is that for many days with the weather conditions of today on average it will rain in 10% of all cases on the next day; no more no less. Whether it will rain or not on one specific date cannot be predicted by giving a probability.

So for the practice of physics the Bayesian view of probabilities is simply pointless, because doesn't tell anything about the outcome real experiments.
 
  • #13
vanhees71 said:
So for the practice of physics the Bayesian view of probabilities is simply pointless, because doesn't tell anything about the outcome real experiments.

I think it goes beyond physics. My background is in applied math and it invariably uses the frequentest interpretation which is basically the same as the ensemble interpretation. To me this Bayesian stuff seems just a play on words as well.

That said, and I can't comment because I didn't take those particular courses, applied Bayesian modelling and inference is widely taught - courses on it were certainly available where I went. I am not convinced however it requires the Bayesian interpretation.

Thanks
Bill
 
  • #14
vanhees71 said:
The Bayesian view, for me, is just a play with words, trying to give a physically meaningful interpretation of probability for a single event. In practice, however, you cannot prove anything about a probabilistic statement with only looking at a single event.

Well, the history of the universe only happens once, so we're stuck with having to reason about singular events, if we are to describe the universe.

More concretely, let's say that we have a theory that predicts that some outcome of an experiment has a 50/50 probability. So you perform the experiment 100 times, say, and find that outcome happens 49 times out 100. So that's pretty good. But logically speaking, how is making the conclusion based on 100 trials any more certain than making a conclusion based on 1 trial, or 10 trials? The outcome, 49 out of 100, is consistent with just about any probability at all. You haven't narrow down the range of probabilities at all. What have you accomplished, then? You've changed your confidence, or belief, that the probability is around 1/2.

Mathematically speaking, the frequentist account of probability is nonsense. Probability 1/2 doesn't mean that something will happen 1/2 of the time, no matter how many experiments you perform. And it's nonsensical to add "...in the limit as the number of trials goes to infinity...", also. There is no guarantee that relative frequencies approach any limit whatsoever.
 
  • #15
bhobba said:
I think it goes beyond physics. My background is in applied math and it invariably uses the frequentest interpretation which is basically the same as the ensemble interpretation. To me this Bayesian stuff seems just a play on words as well.

The frequentist interpretation really doesn't make any sense, to me. As a statement about ensembles, it doesn't make any sense, either. If you perform an experiment, such as flipping a coin, there is no guarantee that the relative frequency approaches anything at all in the limit as the number of coin tosses goes to infinity. Furthermore, since we don't really ever do things infinitely often, then what can you conclude, as a frequentist, from 10 trials of something. Or 100? Or 1000? You can certainly dutifully write down the frequency, but every time you do another trial, than number is going to change, by a tiny amount. Is the probability changing every time you perform the experiment?
 
  • #16
stevendaryl said:
The frequentist interpretation really doesn't make any sense, to me. As a statement about ensembles, it doesn't make any sense, either. If you perform an experiment, such as flipping a coin, there is no guarantee that the relative frequency approaches anything at all in the limit as the number of coin tosses goes to infinity. Furthermore, since we don't really ever do things infinitely often, then what can you conclude, as a frequentist, from 10 trials of something. Or 100? Or 1000? You can certainly dutifully write down the frequency, but every time you do another trial, than number is going to change, by a tiny amount. Is the probability changing every time you perform the experiment?

In practice, people who claim to be doing "frequentist probability" use "confidence intervals". So if you perform an experiment 100 times, and you get a particular outcome 49 times, then you can say something like: The probability is 49% +/- E, where E is a confidence interval. But it isn't really true. The "true" probability could be 99%. Or the "true" probability could be 1%. You could have just had a weird streak of luck. The choice of E is pretty much ad hoc.
 
  • #17
John Baez gives a discussion of Bayesianism here:
http://math.ucr.edu/home/baez/bayes.html

Here's a snippet:

It's not at all easy to define the concept of probability. If you ask most people, a coin has probability 1/2 to land heads up if when you flip it a large number of times, it lands heads up close to half the time. But this is fatally vague!

After all what counts as a "large number" of times? And what does "close to half" mean? If we don't define these concepts precisely, the above definition is useless for actually deciding when a coin has probability 1/2 to land heads up!

Say we start flipping a coin and it keeps landing heads up, as in the play Rosencrantz and Guildenstern are Dead by Tom Stoppard. How many times does it need to land heads up before we decide that this is not happening with probability 1/2? Five? Ten? A thousand? A million?

This question has no good answer. There's no definite point at which we become sure the probability is something other than 1/2. Instead, we gradually become convinced that the probability is higher. It seems ever more likely that something is amiss. But, at any point we could turn out to be wrong. We could have been the victims of an improbable fluke.

Note the words "likely" and "improbable". We're starting to use concepts from probability theory - and yet we are in the middle of trying to define probability! Very odd. Suspiciously circular.

Some people try to get around this as follows. They say the coin has probability 1/2 of landing heads up if over an infinite number of flips it lands heads up half the time. There's one big problem, though: this criterion is useless in practice, because we can never flip a coin an infinite number of times!

Ultimately, one has to face the fact that probability cannot be usefully defined in terms of the frequency of occurence of some event over a large (or infinite) number of trials. In the jargon of probability theory, the frequentist interpretation of probability is wrong.

Note: I'm not saying probability has nothing to do with frequency. Indeed, they're deeply related! All I'm saying is that we can't usefully define probability solely in terms of frequency.
 
  • #18
First, I don't know enough QM to have any opinion on interpretations of QM, but I do use Bayesian statistics in other things (e.g. analysis of medical tests)

vanhees71 said:
Well, maybe I'm just too much biased by my training as a physicist to make sense of the whole Baysian interpretation of probabilities. ...

So for the practice of physics the Bayesian view of probabilities is simply pointless, because doesn't tell anything about the outcome real experiments.
You should really invest a little time into it. The Bayesian approach to probaility is more in line with the scientific method than the frequentist approach.

In the scientific method you formulate a hypothesis, then you acquire data, then you use that data to decide to keep or reject your hypothesis. In other words, you want to determine the likelyhood of the hypothesis given the data, which is exactly what Bayesian statistics calculate. Unfortunately, frequentist statistical tests simply don't measure that. Instead they calculate the likelyhood of the data given the hypothesis.

I think that the big problem with Bayesian statistics right now is the lack of standardized tests. If you say "my t-test was significant with p=0.01" then everyone understands what mathematical test you ran on your data and what you got. There is no corresponding "Bayesian t-test" that you can simply report and expect everyone to know what you did.

Most likely, your preference for frequentist statistics is simply a matter of familiarity, born of the fact that the tools are well-developed and commonly-used. This seems to be the case for bhobba also.
bhobba said:
My background is in applied math and it invariably uses the frequentest interpretation which is basically the same as the ensemble interpretation.
 
  • #19
DaleSpam said:
You should really invest a little time into it. The Bayesian approach to probaility is more in line with the scientific method than the frequentist approach.

actually are merging. (frequentist and bayesian)

"Efron also compares more recent statistical theories such as frequentism to Bayes' theorem, and looks at the newly proposed fusion of Bayes' and frequentist ideas in Empirical Bayes. Frequentism has dominated for a century and does not use prior information, considering future behavior instead"

Read more at: http://phys.org/news/2013-06-bayesian-statistics-theorem-caution.html#jCp
 
  • #20
DaleSpam said:
First, I don't know enough QM to have any opinion on interpretations of QM, but I do use Bayesian statistics in other things (e.g. analysis of medical tests)

You should really invest a little time into it. The Bayesian approach to probaility is more in line with the scientific method than the frequentist approach.

I don't think that the frequentist interpretation of probability can be taken seriously, for the reasons that John Baez gives in the passage I quoted. On the other hand, a purely subjective notion of probability doesn't seem like the whole story, either.

For example, one could model a coin flip by using an unknown parameter [itex]h[/itex] reflecting the probability of getting a "heads". One could start off with the completely unknown probability distribution on [itex]h[/itex]: it could be anything between 0 and 1. Then you flip the coin a few times, and you use Bayes' theorem to get an adjusted probability distribution on the parameter [itex]h[/itex]. For example, if I flip twice, and get 1 head and 1 tail, then the adjusted probability distribution is [itex]P(h) = 6h (1-h)[/itex], which has a maximum at [itex]h=\frac{1}{2}[/itex].

The weird thing here is that you have probability appearing as an unknown parameter, [itex]h[/itex], and you also have it appearing as a subjective likelihood of that parameter. It doesn't make sense to me that it could all be subjective probability, because how can there be an unknown subjective probability [itex]h[/itex]?
 
  • #21
stevendaryl said:
The frequentist interpretation really doesn't make any sense, to me. As a statement about ensembles, it doesn't make any sense, either. If you perform an experiment, such as flipping a coin, there is no guarantee that the relative frequency approaches anything at all in the limit as the number of coin tosses goes to infinity. Furthermore, since we don't really ever do things infinitely often, then what can you conclude, as a frequentist, from 10 trials of something. Or 100? Or 1000? You can certainly dutifully write down the frequency, but every time you do another trial, than number is going to change, by a tiny amount. Is the probability changing every time you perform the experiment?

The law of large numbers is rigorously provable from the axioms pf probability.

What it says is if a trial (experiment or whatever) is repeated a large number of times, independently under identical conditions, then the proportion of times that any specified outcome occurs approximately equals the probability of the event's occurrence on any particular trial; the larger the number of repetitions, the better the approximation tends to be.

This guarantees a sufficiently large, but finite, number of trials exists (ie an ensemble) that for all practical purposes contains the outcomes in proportion to its probability.

It seems pretty straightforward to me, but each to his own I suppose. This is applied math after all. I remember when I was doing my degree I used to get upset at careless stuff like treating dx as a small first order quantity - which of course it isn't - but things are often simpler doing that. Still the criticisms I raised then are perfectly valid and its only in a rigorous treatment they disappear - but things become a lot more difficult. If that's what appeals be my quest - I like to think I have come to terms with such things these days. As one of my statistics professors said to me - and I think it was the final straw that cured me of this sort of stuff - I can show you books where all the questions you ask are fully answered - but you wouldn't read them. He then gave me this deep tome on the theory of statistical inference - and guess what - he was right.

Thanks
Bill
 
  • #22
DaleSpam said:
Most likely, your preference for frequentist statistics is simply a matter of familiarity, born of the fact that the tools are well-developed and commonly-used. This seems to be the case for bhobba also.

That's true.

I too did a fair amount of statistics in my degree deriving and using stuff like the student t distribution, degrees of freedom etc. I always found the frequentest interpretation more than adequate.

That's not to say the Bayesian view is not valid - it is - its just I never found the need to move to that view.

Thanks
Bill
 
  • #23
bhobba said:
The law of large numbers is rigorously provable from the axioms pf probability.

What it says is if a trial (experiment or whatever) is repeated a large number of times, independently under identical conditions, then the proportion of times that any specified outcome occurs approximately equals the probability of the event's occurrence on any particular trial; the larger the number of repetitions, the better the approximation tends to be.

But the meaning of "tends to" is the part that makes no sense, under a frequentist account. What does that mean? It is possible, when flipping coins, to have a run of 1,000 flips in a row that are all heads. It is possible to have a run of 1,000,000 flips in a row with all heads. So what does this "tends to" mean? Well, you can say that such a run of heads is unlikely, but according to what meaning of "unlikely"?

This guarantees a sufficiently large, but finite, number of trials exists (ie an ensemble) that for all practical purposes contains the outcomes in proportion to its probability.

It doesn't guarantee anything. You can calculate a number, the likelihood that the relative frequency will differ from the probability by more than some specific amount. But what's the meaning of that number? It can't be given a frequentist interpretation as a probability.
 
  • #24
stevendaryl said:
But the meaning of "tends to" is the part that makes no sense, under a frequentist account. What does that mean? It is possible, when flipping coins, to have a run of 1,000 flips in a row that are all heads. It is possible to have a run of 1,000,000 flips in a row with all heads. So what does this "tends to" mean? Well, you can say that such a run of heads is unlikely, but according to what meaning of "unlikely"?

The meaning of such things lies in a rigorous development of probability. That's how it is proved and rests on ideas like almost surely convergence and convergence in probability.

You are putting the cart before the horse. In the frequentest interpretation you have an ensemble that is the EXPECTED outcome of a very large number of trials - and that's what the law of large numbers converges to. Sure - you can flip any number of heads - but that is not the expected value - which is for a large number half and half. Then one imagines a trial as picking a random element from that ensemble - you can pick any element of that ensemble every time - but the ensemble contains the objects in the correct proportion.

I agree these ideas are subtle and many great mathematicians such as Kolmorgorov wrote difficult tomes putting all this stuff on a firm basis eg the strong law of large numbers. But on a firm basis they most certainly are.

If you really want to investigate it without getting into some pretty hairy and advanced pure math then Fellers classic is a good place to start:
http://ruangbacafmipa.staff.ub.ac.id/files/2012/02/An-Introduction-to-probability-Theory-by-William-Feller.pdf

As you will see there is a lot of stuff that leads up to the proof of the law of large numbers and in volume 1 Feller does not give the proof of the very important Strong Law Of Large Numbers - he only gives the proof of the Weak Law - you have to go to volume 2 for that - and the level of math in that volume rises quite a bit.

Feller discusses the issues you raise and it is very subtle indeed - possibly even more subtle than you realize. However if that's what interests you then Feller is a good place to start.

Just as an aside I set myself the task of working through both volumes - bought them both. Got through volume 1 but volume 2 was a bit too tough and never did finish it.

Thanks
Bill
 
Last edited by a moderator:
  • #25
bhobba said:
I always found the frequentest interpretation more than adequate.
Yes, I have also, particularly with respect to the wide variety of powerful software with pre-packaged standard statistical tests. Personally, I think that the hard-core Bayesians need to spend less time promoting their viewpoint and more time developing and standardizing their tools.
 
  • #26
DaleSpam said:
Yes, I have also, particularly with respect to the wide variety of powerful software with pre-packaged standard statistical tests. Personally, I think that the hard-core Bayesians need to spend less time promoting their viewpoint and more time developing and standardizing their tools.

I am a bit more sanguine on it personally. I think either view is valid and the choice is a personal preference, its just, like you correctly said, the early probability courses at university, and certainly at high school, use the frequentest version almost exclusively so its a bit of a paradigm shift to move away from it.

I think most people are like me and want a real good reason to make the effort.

Thanks
Bill
 
  • #27
DaleSpam said:
First, I don't know enough QM to have any opinion on interpretations of QM, but I do use Bayesian statistics in other things (e.g. analysis of medical tests)

You should really invest a little time into it. The Bayesian approach to probaility is more in line with the scientific method than the frequentist approach.

In the scientific method you formulate a hypothesis, then you acquire data, then you use that data to decide to keep or reject your hypothesis. In other words, you want to determine the likelyhood of the hypothesis given the data, which is exactly what Bayesian statistics calculate. Unfortunately, frequentist statistical tests simply don't measure that. Instead they calculate the likelyhood of the data given the hypothesis.

I think that the big problem with Bayesian statistics right now is the lack of standardized tests. If you say "my t-test was significant with p=0.01" then everyone understands what mathematical test you ran on your data and what you got. There is no corresponding "Bayesian t-test" that you can simply report and expect everyone to know what you did.

Most likely, your preference for frequentist statistics is simply a matter of familiarity, born of the fact that the tools are well-developed and commonly-used. This seems to be the case for bhobba also.

I admit, I don't understand the Bayesian interpretation of probabilities. Particularly the subjectivity makes it highly suspicious for me.

In the natural sciences (and hopefully also in medicine and the social sciences) to the contrary, one has to try to make statements with the "least prejudice", given the (usually incomplete) information. One method is the maximum-entropy principle (based on information theory developed by Shannon and applied to (quantum) statistical physics by Jaynes). Such probabilities are subjective in the sense that you adapt your probability distribution to the (incomplete) information you have, but due to the maximum-entropy principle you use the distribution of least prejudice (in the sense of Shannon-Jaynes information theory).

In quantum theory, at least in the minimal statistical interpretation, probabilities are even "more objective" in the case of pure states. If you have prepared a system in a pure state, this means (from the theoretical point of view) that you have determined a complete set of compatible observables due to this preparation. Then quantum theory tells you precisely which probabilities for outcomes of further measurements you have to use (using the Heisenberg picture of time evolution for simplicity): The associated state of the system is given by the common eigenvector of the representing self-adjoint commuting operators of the complete set of compatible observables, which is uniquely defined up to a phase factor. The probability to measure any observable of the system at a later time is then given by the Born rule. There is no choice anymore which probability distribution you should associate with the situation. BTW. It's compatible with the maximum-entropy principle since given the values of the complete set of compatible observables it leads to the pure state, represented by the corresponding eigenvector.

For me the subjectivistic view of probabilities as seems to be used in Bayesian statistics is not scientific at all. Indeed, natural science is empirical, and you build theories which predict the outcome of measurements. These predictions have to be experimentally tested. Any experiment must be able to be reproducible precisely enough such that you can get "high enough statistics" to check a hypothesis quantitatively, i.e., to get the statistical significance of your measurement. Here, for simplicity, I assume an ideal experiment which has only negligible systematical errors.

All this uses the frequentist interpretation of probabilities, which is based on the law of large numbers, as bhobba already mentioned.

Perhaps it would help me to understand the Bayesian view, if you could explain how to test a probilistic theoretical statement empirically from this point of view. Is there a good book for physicists to understand the Bayesian point of view better?
 
  • #28
vanhees71 said:
For me the subjectivistic view of probabilities as seems to be used in Bayesian statistics is not scientific at all.

It scientific - just based on a different set of assumptions that's all.

What you have is three values of logic - true, false, and don't know - but don't know is assigned some kind of value indicating how close it is to true or false ie its plausibility. Its an extension of formal logic.

This value is assumed to obey some reasonable axioms - usually the so called Cox axioms. One of the key theorems is they imply the Kolmogorov axioms - with a minor difference - probabilities are finitely but not countably additive so one needs a slight further assumption to handle the infinite case.

As I indicated before the Kolmogrov axioms imply the frequentest interpretation via the law of large numbers which is provable in those axioms. What it says is, via some kind of probabilistic convergence, as the number of trials increases the number of each possible outcome approaches its expected number. Probabilistic convergence means the probability of it being that value gets higher and higher as it get closer and closer. This means, in the case of the law of large numbers, it is infinitesimally close 1 for a suitably large number. Now just like is done in many areas of applied math it is taken for all practical purposes to be 1 if it is infinitesimally close to it ie a dead cert it will have the proportion of outcomes the same as the probability.

What Stevendaryl seems to be concerned about is it is only infinitesimally close to it. As I say - each to their own I suppose - but I suspect virtually everyone that applies probability will accept it.

Thanks
Bill
 
  • #29
bhobba said:
The meaning of such things lies in a rigorous development of probability. That's how it is proved and rests on ideas like almost surely convergence and convergence in probability.

That's my point---frequentism can't be the basis of a rigorous development of probability. It can be an intuitive way to think about probability, but it doesn't actually work as a definition of probability.

You are putting the cart before the horse. In the frequentest interpretation you have an ensemble that is the EXPECTED outcome of a very large number of trials - and that's what the law of large numbers converges to.

Expected value is defined in terms of probability. So you can't use frequencies to be your definition of probability.
 
  • #30
bhobba said:
What Stevendaryl seems to be concerned about is it is only infinitesimally close to it. As I say - each to their own I suppose - but I suspect virtually everyone that applies probability will accept it.

"Infinitesimally close" is defined in terms of probability, but a notion of probability that is NOT relative frequency. So frequencies cannot be the core definition of probability.

I don't have any complaints with the use of frequencies as "frequentists" use them. It works for practical purposes, but it's not an actually consistent theory. In that sense, frequentism is sort of like the Copenhagen or "collapse" version of quantum mechanics. It's a set of rules for using probabilities, and it works well enough, but doesn't actually count as a rigorous theory.
 
  • #31
stevendaryl said:
"Infinitesimally close" is defined in terms of probability, but a notion of probability that is NOT relative frequency. So frequencies cannot be the core definition of probability.

I don't have any complaints with the use of frequencies as "frequentists" use them. It works for practical purposes, but it's not an actually consistent theory. In that sense, frequentism is sort of like the Copenhagen or "collapse" version of quantum mechanics. It's a set of rules for using probabilities, and it works well enough, but doesn't actually count as a rigorous theory.

Yes - by the Kolmogorov axioms. One of those axioms is the entire event space has probability 1. This means if an element of the event space is 1 it can be considered the entire event space. The law of large numbers shows for large n the outcomes are in proportion to the probabilities with probability so close to 1 it is for all practical purposes equal to it ie it is the entire event space.

If you don't consider that consistent then there are many areas in applied math where infinitesimal quantities are ignored so I guess you have trouble with those as well.

Thanks
Bill
 
  • #32
stevendaryl said:
That's my point---frequentism can't be the basis of a rigorous development of probability.

It isn't - the Kolmogorov axioms are. And that is what rigorous tretments use - not the frequentest interpretation. But the law of large numbers show for all practical purposes the frequentest interpretation is equivalent to it.

However I am taken back to what I was warned about all those years ago - you wouldn't read the tomes based on it - and in my experience that is very true. Even Fellers book, especially volume 2, is hard going, and believe me there are worse than that about.

Thanks
Bill
 
  • #33
Bayesianism vs Falsifiability

vanhees71 said:
Perhaps it would help me to understand the Bayesian view, if you could explain how to test a probilistic theoretical statement empirically from this point of view. Is there a good book for physicists to understand the Bayesian point of view better?

To me, Bayesianism provides a slightly different point of view about what it means to do science than the Karl Popper view that places "falsifiability" at the core.

In the Karl Popper view, we create a theory, we use the theory to make a prediction, we perform an experiment to test the prediction, and we discard the theory if the prediction is not confirmed. That's a nice, neat way of thinking about science, but it's an over-simplification, and it's incomplete. Why do I say that?

First, why is it an over-simplification? Because there is no way that a single experiment can ever falsify a theory. Whatever outcome happens from an experiment is capable of having multiple explanations. Some of those explanations imply that the theory used to make the prediction is simply wrong, and some do not. For example, if there is a glitch in a piece of equipment, you can't hold that against the theory. Also, every experiment that attempts to test a theory relies on interpretations that go beyond the theory. You don't directly measure "electron spin", for example, you measure the deflection of the electron in a magnetic field. So you need to know that there are no other forces at work on the electron that might deflect it, other than its magnetic moment (and you also need the theory connecting the electron's spin with its magnetic moment). So when a theory fails to make a correction, the problem could be in the theory, or it could be in the equipment, or it could be some random glitch, or it could be in the theory used to interpret the experimental result, or whatever. So logically, you can never absolutely falsify a theory. It gets even worse when the theories themselves are probabilistic. If the theory predicts something with 50% probability, and you observe that it has happened 49% of the time, has the theory been falsified, or not? There is no definite way to say.

Second, why do I say that falsifiability is incomplete? It doesn't provide any basis for decision-making. You're building a rocket, say, and you want to know how to design it so that it functions as you would like it to. Obviously, you want to use the best understanding of physics in the design of the rocket, but what does "best" mean? At any given time, there are infinitely many theories that have not yet been falsified. How do you pick out one as the "current best theory"? You might say that that's not science, that's engineering, but the separation is not that clear-cut, because, as I said, you need to use engineering in designing experimental equipment, and you have to use theory to interpret experimental results. You have to pick a "current best theory" or "current best engineering practice" in order to do experiments to test theories.

So how does Bayesianism come to the rescue? Well, nobody really uses Bayesianism in its full glory, because it's mathematically intractable, but it provides a model for how to do science that allows us to see what's actually done as a pragmatic short-cut.

In the Bayesian view, nothing besides pure mathematical claims is ever proved true or false. Instead, claims have likelihoods, and performing an experiment allows you to adjust those likelihoods.

So rather than saying that a theory is falsified by an experiment, the Bayesian would say that the theory's likelihood is decreased. And that's the way it really works. There was no single experiment that falsified Newtonian mechanics. Instead, there was a succession of experiments that cast more and more doubt on it. There was never a point where Newtonian mechanics was impossible to believe, it's just that at some point, the likelihood of Special Relativity and quantum mechanics rose to be higher (and by today, significantly higher) than Newtonian mechanics.

The other benefit, at least in principle, if not in practice, for Bayesianism is that it actually gives us a basis for making decisions about things like how to design rockets, even when we don't know for certain what theory applies. What you can do is figure out what you want to accomplish (get a man safely on the moon, for example), and try to maximize the likelihood of that outcome. If there are competing theories, then you include ALL of them in the calculation of likelihood. Mathematically, if [itex]O[/itex] is the desired outcome, and [itex]E[/itex] is the engineering approach to achieving it, and [itex]T_1, T_2, ...[/itex] are competing theories, then

[itex]P(O | E) = \sum_i P(T_i) P(O | E, T_i)[/itex]

You don't have to know for certain what theory is true to make a decision.
 
  • #34
bhobba said:
It isn't - the Kolmogorov axioms are. And that is what rigorous tretments use - not the frequentest interpretation. But the law of large numbers show for all practical purposes the frequentest interpretation is equivalent to it.

However I am taken back to what I was warned about all those years ago - you wouldn't read the tomes based on it - and in my experience that is very true. Even Fellers book, especially volume 2, is hard going, and believe me there are worse than that about.

Thanks
Bill

Okay, I guess if by "frequentism" you mean a particular methodology for using probabilities, then I don't have any big problems with it. But if it's supposed to explain the meaning of probabilities, I don't think it can actually do that, because you have to already have a notion of probability in order to connect relative frequencies to probabilities.
 
  • #35
stevendaryl said:
Okay, I guess if by "frequentism" you mean a particular methodology for using probabilities, then I don't have any big problems with it. But if it's supposed to explain the meaning of probabilities, I don't think it can actually do that, because you have to already have a notion of probability in order to connect relative frequencies to probabilities.

In modern times, just like many areas of mathematics, probability is defined in terms of axioms that are not given any meaning. Have a look at the Kolmogorov axioms - there is no meaning assigned to it whatsoever. The frequentest interpretation is one way to give it meaning - Baysian is another. This is the great strength of the formalist approach - a proof applies to however you relate it to stuff out there. Its great weakness is without the help of the intuition of an actual application you have to derive everything formally.

I agree justifying the frequentest interpretation without the Kolmogorov axioms leads to problems such as circular arguments. But then again I don't know of any book on probability that does that - all I have ever seen start with the Kolmogorov axioms and show, with varying degrees of rigor, and actually proving the key theorems, that the frequentest view follows from it.

Thanks
Bill
 

Similar threads

  • Quantum Interpretations and Foundations
Replies
9
Views
2K
  • Quantum Interpretations and Foundations
2
Replies
41
Views
4K
  • Quantum Interpretations and Foundations
Replies
1
Views
344
  • Poll
  • Quantum Interpretations and Foundations
Replies
10
Views
358
  • Quantum Interpretations and Foundations
Replies
33
Views
3K
  • Quantum Interpretations and Foundations
Replies
1
Views
494
  • Quantum Interpretations and Foundations
7
Replies
223
Views
6K
  • Quantum Interpretations and Foundations
11
Replies
376
Views
11K
  • Quantum Interpretations and Foundations
Replies
13
Views
824
  • Quantum Interpretations and Foundations
Replies
27
Views
1K
Back
Top