You are on page 1of 3

INTRODUCTION

A probability distribution assigns a probability to each


measurable subset of the possible outcomes of a random
experiment, survey, or procedure of statistical inference.
Examples are found in experiments whose sample space is nonnumerical, where the distribution would be a categorical
distribution; experiments whose sample space is encoded by
discrete random variables, where the distribution can be
specified by a probability mass function; and experiments with
sample spaces encoded by continuous random variables, where
the distribution can be specified by a probability density
function. More complex experiments, such as those involving
stochastic processes defined in continuous time, may demand
the use of more general probability measures.
A probability distribution can either be univariate or
multivariate. A univariate distribution gives the probabilities of
a single random variable taking on various alternative values; a
multivariate distribution (a joint probability distribution) gives
the probabilities of a random vectora set of two or more
random variablestaking on various combinations of values.
Important and commonly encountered univariate probability
distributions include the binomial distribution, the
hypergeometric distribution, and the normal distribution. The
multivariate normal distribution is a commonly encountered
multivariate distribution.
The concept of the probability distribution and the random
variables which they describe underlies the mathematical
discipline of probability theory, and the science of statistics.
There is spread or variability in almost any value that can be
measured in a population (e.g. height of people, durability of a
metal, sales growth, traffic flow, etc.); almost all measurements
are made with some intrinsic error; in physics many processes
are described probabilistically,from the kinetic properties of
gases to the quantum mechanical description of fundamental
particles. For these and many other reasons, simple numbers
are often inadequate for describing a quantity, while probability

distributions are often more appropriate.As a more specific


example of an application, the cache language models and
other statistical language models used in natural language
processing to assign probabilities to the occurrence of
particular words and word sequences do so by means of
probability distributions.
In probability theory, the normal (or Gaussian) distribution is a
very common continuous probability distribution. Normal
distributions are important in statistics and are often used in
the natural and social sciences to represent real-valued random
variables whose distributions are not known.The normal
distribution is remarkably useful because of the central limit
theorem. In its most general form, under mild conditions, it
states that averages of random variables independently drawn
from independent distributions are normally distributed.
Physical quantities that are expected to be the sum of many
independent processes (such as measurement errors) often
have distributions that are nearly normal.[3] Moreover, many
results and methods (such as propagation of uncertainty and
least squares parameter fitting) can be derived analytically in
explicit form when the relevant variables are normally
distributed.The normal distribution is sometimes informally
called the bell curve. However, many other distributions are
bell-shaped (such as Cauchy's, Student's, and logistic). The
terms Gaussian function and Gaussian bell curve are also
ambiguous because they sometimes refer to multiples of the
normal distribution that cannot be directly interpreted in terms
of probabilities.
The example of normal distribution in real life is to determine
the height of 2000 male football players and to determine the
sales receipts for 1000 customers at a grocery store. We can
also determine the size of things produced by machines by
applying the normal distribution.Besides,we can also sove the
errors in measurements and check the blood pressure and
marks on a test.

The binomial distribution with parameters n and p is the


discrete probability distribution of the number of successes in a
sequence of n independent yes/no experiments, each of which
yields success with probability p. A success/failure experiment
is also called a Bernoulli experiment or Bernoulli trial; when n =
1, the binomial distribution is a Bernoulli distribution. The
binomial distribution is the basis for the popular binomial test of
statistical significance.The binomial distribution is frequently
used to model the number of successes in a sample of size n
drawn with replacement from a population of size N. If the
sampling is carried out without replacement, the draws are not
independent and so the resulting distribution is a
hypergeometric distribution, not a binomial one. However, for N
much larger than n, the binomial distribution is a good
approximation, and widely used.
For example, if a new drug is introduced to cure a disease, it
either cures the disease (its successful) or it doesnt cure the
disease (its a failure). If you purchase a lottery ticket, youre
either going to win money, or you arent. Basically, anything
you can think of that can only be a success or a failure can be
represented by a binomial distribution. To find the number of
machines that last longer than T hours of operation without
failure.

You might also like