Probability: Difference between revisions
imported>Aleksander Halicz m (→External links: math workgroup) |
imported>David Tribe |
||
Line 23: | Line 23: | ||
== Example of the | == Example of the Bayesian viewpoint == | ||
We are given a die, and no information about it. We then set up 6 propositions: "The outcome will be 1", "the outcome will be 2", etc. From the principle of maximum entropy, we assume that all 6 outcomes are equally likely. Our state of knowledge as to the outcome is thus best modeled by putting equal "weights" to all 6 alternative propositions, i.e. 1/6 of our total to each. This is then our prior probability distribution. | We are given a die, and no information about it. We then set up 6 propositions: "The outcome will be 1", "the outcome will be 2", etc. From the principle of maximum entropy, we assume that all 6 outcomes are equally likely. Our state of knowledge as to the outcome is thus best modeled by putting equal "weights" to all 6 alternative propositions, i.e. 1/6 of our total to each. This is then our prior probability distribution. | ||
We will continually watch the outcomes, and continually redistribute our "weights" according to the results we obtain from throwing the die. | We will continually watch the outcomes, and continually redistribute our "weights" according to the results we obtain from throwing the die. | ||
== Example of the objective viewpoint == | == Example of the objective viewpoint == |
Revision as of 04:56, 10 February 2007
Probability
Probability theory is a branch of mathematics.
Like algebra, geometry and other parts of mathematics, probability theory has it's origins in the natural world. Humans routinely deal with incomplete and/or uncertain information in daily life, in decisions such as crossing the road ("will this approaching car respect the red light"), eating food ("am I certain this food is not contaminated?"), and so on. Probability theory is a mathematical tool intended to formalize this ubiquitous mental process. The probability concept is a part of this theory, and is intended to formalize uncertainty.
There are two basic ways to think about the probability concept:
- Subjective (Bayesian) probability.
- Objective probability.
The different approaches are largely pedagogical, as some people find one approach or the other much easier.
Bayesian probability
In this approach the probabilities represent a state of knowledge. One starts with a set of propositions and all the available information. Using common sense based formal and/or informal methods one assigns "weights" to the propositions, generating a so called "prior probability distribution". As more information comes in, one combines this "prior" with the new data and obtain a new "posterior" probability distribution, which represents an updated state of knowledge - each probability is a number that describes how much faith one should presently allocate to its associated proposition.
Objective probability
In this approach one views probabilities as "propensities" of the actual system under study - f.i. a fair coin will have a "propensity" to show heads 50% of the time. This approach is more restrictive than the Bayesian interpretation: F.i. there is no way to assign a probability as to whether or not there exists life in the Andromeda galaxy this way, since no "propensities" have been measured.
Example of the Bayesian viewpoint
We are given a die, and no information about it. We then set up 6 propositions: "The outcome will be 1", "the outcome will be 2", etc. From the principle of maximum entropy, we assume that all 6 outcomes are equally likely. Our state of knowledge as to the outcome is thus best modeled by putting equal "weights" to all 6 alternative propositions, i.e. 1/6 of our total to each. This is then our prior probability distribution.
We will continually watch the outcomes, and continually redistribute our "weights" according to the results we obtain from throwing the die.
Example of the objective viewpoint
We are given a die, and a list of it's measured (or theoretical) "propensities", i.e. probabilities for each possible outcome. We then use this information to calculate the "propensities"/probabilities of certain outcomes. If our (empirical) results seem improbable, we may decide to do experiments to re-measure the "propensities".
More technical information
- Bayes theorem
- principle of maximum entropy
- Probability distributions
- Kolmogorov's axioms
- probability theory
Links
External links
- Intros