Relationship between a posterior distribution and the LLN

In summary: I see. In that case, is there a way to relate Bayes' theorem to the LLN without explicitly mentioning the coin example?Yes, there is. Bayes' theorem can be used to calculate the expected value of a particular event or outcome, regardless of the coin example.Yes, there is. Bayes' theorem can be used to calculate the expected value of a particular event or outcome, regardless of the coin example.
  • #1
scinoob
17
0
Hello everybody. This is my first post here and I hope I'm not asking a question that's been addressed already (I did try to use the search function, but couldn't find what I'm looking for).

Both the Bayes theorem and the law of large numbers are mathematical theorems derived from Kolmogorov's axioms. I've been thinking of a way to relate the two and I'm not sure how that can be done.

Let's say I have a coin with an unknown bias towards heads Θ. I start with a uniform prior distribution of Θ (between 0 and 1 obviously) and keep flipping the coin while updating the prior using Bayes' theorem. Let's say after 300 flips the posterior looks like a relatively peaked unimodal distribution centered at 0.3. Then I decide to flip the coin another, say, 10k times without updating the posterior anymore and somebody asks the question "approximately what percentage of heads do you expect to see after the 10k flips?" So, what I am likely going to do (at least according to the standard practice) is calculate the expected value of a single flip using the posterior distribution I obtained after the 300 flips. If I then go ahead and flip the coin 10k times, can I expect that the percentage will be close to the value I obtained in the analytic calculation? If I keep flipping it, can I expect that the % of heads will converge to the expected value by appealing to the LLN?

If the answer is 'no', then is there another way to relate Bayes' theorem to the LLN, say, using my coin example in particular?
 
Physics news on Phys.org
  • #2
scinoob said:
Let's say after 300 flips the posterior looks like a relatively peaked unimodal distribution centered at 0.3.

Bayesians vary in how they interpret the bayesian approach to problems. One common way to think about the posterior distribution in your problem is that your particular coin was picked at from a population of coins with various probabilities of landing heads and that the posterior distribution gives the probabilities of picking such a coin. An expected value calculated using the posterior distribution is the expected number of heads over that entire population of coins, not the expected value of heads from tossing your particular coin repeatedly. So when you say that you continue to toss the coin 10k more times, you must be specific about whether you are tossing the same coin 10k times or whether you are imagining each of the 10k tosses of a coin to consist of two steps: a) Pick the coin at random from the population of all coins b) toss the coin you picked.
 
  • #3
Stephen Tashi said:
Bayesians vary in how they interpret the bayesian approach to problems. One common way to think about the posterior distribution in your problem is that your particular coin was picked at from a population of coins with various probabilities of landing heads and that the posterior distribution gives the probabilities of picking such a coin. An expected value calculated using the posterior distribution is the expected number of heads over that entire population of coins, not the expected value of heads from tossing your particular coin repeatedly. So when you say that you continue to toss the coin 10k more times, you must be specific about whether you are tossing the same coin 10k times or whether you are imagining each of the 10k tosses of a coin to consist of two steps: a) Pick the coin at random from the population of all coins b) toss the coin you picked.

Good point. In this case I mean tossing the same coin 10k times. In other words, it's the same coin with some unknown bias θ. The posterior distribution represents the degree of belief in each value for θ.
 
  • #4
If I keep flipping it, can I expect that the % of heads will converge to the expected value by appealing to the LLN?
No, you can't justify that by the CLT.
 
  • #5
Stephen Tashi said:
No, you can't justify that by the CLT.

I'm afraid I have to disagree here. I think this is exactly what the CLT says. It will converge to the true mean of the biased coin. It will not converge to the mean of the 300 samples unless that happens to have 0 error.
 
  • #6
But the true mean of the particular biased coin is not the mean calculated from the posterior distribution, which considers coins with a variety of biases in the calculation.

(Also "will converge" has a very technical definition - different than the definition of convergence used in calculus for the convergence of sequences or limits of functions.)
 
Last edited:
  • #7
Stephen Tashi said:
But the true mean of the particular biased coin is not the mean calculated from the posterior distribution, which considers coins with a variety of biases in the calculation.

(Also "will converge" has a very technical definition - different than the definition of convergence used in calculus for the convergence of sequences or limits of functions.)

Oh. I missed that there were multiple coins with different biases. Then you are right.
 

Related to Relationship between a posterior distribution and the LLN

1. What is a posterior distribution in statistics?

A posterior distribution is a probability distribution that represents our updated beliefs about a parameter or unknown quantity after taking into account new information or data. It is calculated using Bayes' theorem and is based on both prior beliefs and observed data.

2. What is the relationship between a posterior distribution and the law of large numbers (LLN)?

The Law of Large Numbers states that as the sample size increases, the sample mean will converge to the true population mean. The posterior distribution takes into account both prior beliefs and observed data, so as more data is collected, the posterior distribution will also converge to the true distribution of the parameter.

3. How does the posterior distribution affect our understanding of a parameter?

The posterior distribution provides a more accurate representation of the true distribution of a parameter by taking into account both prior beliefs and observed data. It allows us to update our understanding of the parameter and make more informed decisions or predictions.

4. Can the posterior distribution be used to make predictions?

Yes, the posterior distribution can be used to make predictions by using it as the basis for a probability distribution for future events. This allows for more accurate and informed predictions based on both prior beliefs and observed data.

5. Are there any limitations to the relationship between a posterior distribution and the LLN?

One limitation is that the LLN assumes that the sample size is large enough for the sample mean to converge to the true population mean. If the sample size is too small, the posterior distribution may not accurately represent the true distribution of the parameter. Additionally, the posterior distribution is only as accurate as the prior beliefs and observed data used to calculate it.

Similar threads

  • Set Theory, Logic, Probability, Statistics
Replies
1
Views
990
  • Set Theory, Logic, Probability, Statistics
Replies
1
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
3
Views
1K
  • Set Theory, Logic, Probability, Statistics
2
Replies
45
Views
3K
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
4K
  • Set Theory, Logic, Probability, Statistics
Replies
4
Views
3K
  • Programming and Computer Science
Replies
2
Views
758
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
3K
  • Precalculus Mathematics Homework Help
Replies
2
Views
870
Back
Top