# Basic Probability

The probability for a given event can be thought of as the ratio of the number of ways that event can happen divided by the number of ways that any possible outcome could happen. If we identify the set of all possible outcomes as the "sample space" and denote it by S, and label the desired event as E, then the probability for event E can be written

In the probability of a throw of a pair of dice, bet on the number 7 since it is the most probable. There are six ways to throw a 7, out of 36 possible outcomes for a throw. The probability is then

The idea of an "event" is a very general one. Suppose you draw five cards from a standard deck of 52 playing cards, and you want to calculate the probability that all five cards are hearts. This desired event brings in the idea of a combination. The number of ways you can pick five hearts, without regard to which hearts or which order, is given by the combination

while the total number of possible outcomes is given by the much larger combination

The same basic probability expression is used, but it takes the form

So drawing a five-card hand of a single selected suit is a rare event with a probability of about one in 2000.

If you want the probabability that any one of a number of disjoint events will occur, the probabilities of the single events can be added. For example, the probability of drawing five cards of any one suit is the sum of four equal probabilities, and four times as likely. In boolean language, if the events are related by a logical OR, then the probabilities add.

If the events are related by a logical AND, the resultant probability is the product of the individual probabilities. If you want the probability of throwing a 7 with a pair of dice AND throwing another 7 on the second throw, then the probability would be the product

The expression for probability must be such that the addition of the probabilities for all events must be 1. Constraining the sum of all the probabilities to be 1 is called "normalization". When you calculate the probability by direct counting processes like those discussed above, then the probabilities are always normalized. But when you develop expressions for the probability of events in nature, you must make sure that your probability expression is normalized.

 The probability distribution function
Index

Statistics concepts

 HyperPhysics*****HyperMath R Nave
Go Back

# Probability Distribution Function

Dealing with basic probability as a discrete counting process is satisfactory if you have reasonably small numbers, like throwing dice or picking cards. But if the number of events is very large, as in the distribution of energy among the molecules of a gas, then the probability can be approximated by a continuous variable so that the methods of calculus can be used.

Using the variable x to represent a possible outcome or event, then in the discrete case the basic framework could be summarized as follows:

 For a particular outcome xi P(xi) = probability that xi will be observed SiP(xi) =1 (normalization)

If you allow the outcome x to take a continuous range of values, then the probability P(x) takes a different character, since to get a finite result for probability, you must sum the probability over a finite range of x. Since x is a continuous variable, this sum takes the form of an integral. A common practice to define a distribution function as a derivative of the probability

The probability of finding outcomes between x=a and x=b can then be expressed as

and the normalization condition is then

For application of probability to physical processes, the use of the distribution function is a very useful strategy. Perhaps the most widely used distribution function in classical physics is the Boltzmann distribution function, which describes the probability of finding particles with an amount of energy E at a given temperature T.

 Examples of distribution functions
 Physical applications of distribution functions
Index

Statistics concepts

 HyperPhysics*****HyperMath R Nave
Go Back