2059: Modified Bayes' Theorem
|Modified Bayes' Theorem|
Title text: Don't forget to add another term for "probability that the Modified Bayes' Theorem is correct."
| This explanation may be incomplete or incorrect: Please edit the explanation below and only mention here why it isn't complete. Do NOT delete this tag too soon.|
If you can address this issue, please edit the page! Thanks.
Bayes' theorem is:
- is the probability that , the hypothesis, is true given observation . This is called the posterior probability.
- is the probability that observation will appear given the truth of hypothesis . This term is often called the likelihood
- is the probability that hypothesis is true before any observations. This is called the prior
- is the probability of the observation X regardless of any hypothesis might have produced it. This term is called the marginal likelihood.
The purpose of Baysian inference is to discover something we want to know (how likely is it that our explanation is correct given the evidence we've seen) by mathematically expressing it in terms of things we can find out: how likely are our observations, how likely is our hypothesis a priori, and how likely are we to see the observations we've seen assuming our hypothesis is true. A bayesian learning system will iterate over available observations, each time using the likelihood of new observations to update its priors (beliefs) with the hope that, after seeing enough data points, the prior and posterior will converge to a single model.
If the modified theorem reverts to the original Bayes' theorem (which makes sense, as a probability one would mean certainty that you are using Bayes' theorem correctly).
If the modified theorem becomes , which says that the belief in your hypothesis is not affected by the result of the observation (which makes sense because you're certain you're misapplying the theorem so the outcome of the calculation shouldn't affect your belief.)
This happens because, if you apply the original theorem, the modified theorem can be rewritten as: . This is the linear-interpolated weighted average of the belief you had before the calculation and the belief you would have if you applied the theorem correctly. This goes smoothly from the not believing your calculation at all, keeping the same belief as before if to changing your belief exactly as Bayes' theorem suggests when .
is the probability that you are using the theorem incorrectly.
As an equation, the rewritten form makes no sense. is strangely self-referential and reduces to the piecewise equation . However, the Modified Bayes Theorem includes an extra variable not listed in the conditioning, so a person with an AI background might understand that Randal was trying to write an expression for updating with knowledge of i.e. , the belief in the hypothesis given the observation and the confidence that you were applying Bayes' theorem correctly , for which the expression makes some intuitive sense.
The title text suggests that an additional term should be added for the probability that the Modified Bayes Theorem is correct. But that's *this* equation, so it would make the formula self-referential. It could also result in an infinite regress -- we'd need another term for the probability that the version with the probability added is correct, and another term for that version, and so on. It's also unclear what the point of using an equation we're not sure of is (although sometimes we can: Newton's Laws are not as correct as the Einstein's Theory of Relativity but they're a reasonable approximation in most circumstances}.
|This transcript is incomplete. Please help editing it! Thanks.|
- Modified Bayes' theorem:
- P(H|X) = P(H) × (1 + P(C) × ( P(X|H)/P(X) - 1 ))
- H: Hypothesis
- X: Observation
- P(H): Prior probability that H is true
- P(X): Prior probability of observing X
- P(C): Probability that you're using Bayesian statistics correctly
add a comment! ⋅ add a topic (use sparingly)! ⋅ refresh comments!