Mathematical explanation - Normal distribution

in #mathematics6 years ago (edited)

Hello friends of steemit continuing with my publications referring to the science of mathematics today I want to make reference in this post about the normal distribution, which is very important for any statistical process since it allows us to transform the variables into approximate values to the study of certain In this publication, you define your concept, its properties, the variables that define it, as well as the central limit theorem, to complement this information you can see at the end of this post some examples of basic exercises where the standard normal distribution is applied.

Normal distribution

In statistics and probability it is called normal distribution, Gaussian distribution or Gaussian distribution, to one of the probability distributions of continuous variable that most frequently appears statistics and probability theory.

The graph of its density function has a flared shape and is symmetric with respect to a certain statistical parameter. This curve is known as Gaussian bell and is the graph of a Gaussian function.

dn.png

The green line corresponds to the standard normal distribution

Probability density function

The importance of this distribution lies in the fact that it allows to model numerous natural, social and psychological phenomena. While the mechanisms that underlie much of this type of phenomena are unknown, due to the enormous amount of uncontrollable variables that intervene in them, the use of the normal model can be justified assuming that each observation is obtained as the sum of a few causes independent

In fact, descriptive statistics only allows describing a phenomenon, without any explanation. For the causal explanation, experimental design is necessary, hence the use of statistics in psychology and sociology is known as a correlational method.

The normal distribution is also important because of its relation to the least squares estimation, one of the simplest and oldest estimation methods.

Some examples of variables associated with natural phenomena that follow the normal model are:

  • morphological characters of individuals such as stature;
  • physiological characters such as the effect of a drug;
  • sociological characters such as the consumption of a certain product
    by the same group of individuals;
  • Psychological characters such as the IQ;
  • noise level in telecommunications;
  • errors made in measuring certain magnitudes;
    etc.

The normal distribution also appears in many areas of the statistic itself. For example, the sample distribution of the sample means is approximately normal, when the distribution of the population from which the sample is drawn is not normal. In addition, the normal distribution maximizes the entropy between all the distributions with known mean and variance, which makes it the natural choice of the underlying distribution to a list of summarized data in terms of sample mean and variance. The normal distribution is the most widespread in statistics and many statistical tests are based on a more or less justified "normality" of the random variable under study.

In probability, the normal distribution appears as the limit of several continuous and discrete probability distributions.

dn1.png

Probability distribution function

Formal definition

The distribution function of the normal distribution is defined as follows:

dn3.png

Therefore, the distribution function of the standard normal is:

dn4.png

This distribution function can be expressed in terms of a special function called error function in the following way:

dn5.png

and the distribution function itself can, therefore, be expressed as follows:

dn6.png

The complement of the distribution function of the standard normal is denoted with frequency Q (x), and is sometimes referred to simply as a Q function, especially in engineering texts. This represents the probability tail of the Gaussian distribution. Other definitions of the Q function are also occasionally used, which are all simple transformations.

The inverse of the distribution function of the standard normal (quantile function) can be expressed in terms of the inverse of the error function:

dn7.png

and the inverse of the distribution function can, therefore, be expressed as:

dn8.png

This quantile function is sometimes called the probit function. There is no elementary primitive for the probit function. This does not mean merely that it is not known, but that the absence of such a function has been proven. There are several exact methods to approximate the quantile function by the normal distribution (see quantile function).

The values Φ (x) can be approximated very accurately by different methods, such as numerical integration, Taylor series, asymptotic series and continuous fractions.

Strict lower and upper limit for the distribution function

For large values of x the distribution function of the standard normal Φ (x) is very close to dn9.png is very close to 0. The elementary limits

dn10.png

in terms of density they are useful.

Using the variable change v = u² / 2, the upper limit is obtained as follows:

dn11.png

Similarly, using the quotient rule,

dn12.png

Generating functions

Moment generating function

The moment generator function is defined as the expectation of e (tX). For a normal distribution, the moment generating function is:

dn13.png

as can be verified by completing the square in the exponent.

Characteristic function

The characteristic function is defined as the expectation of e (itX), where i is the imaginary unit. In this way, the characteristic function is obtained by replacing t by it in the generatrix function of moments. For a normal distribution, the characteristic function is

dn14.png

Properties

Some properties of the normal distribution are the following:

dn15.png

Probability distribution around the mean in a distribution

1 . It is symmetric about its average dn17.png
2 . Fashion and the median are both equal to the average, dn17.png
3 . The turning points of the curve are given for dn19.png and dn20.png
4 . Distribution of probability in an environment of the mean:

1 . in the interval dn21.png it is included, approximately, 68.26% of the distribution.

2 . in the interval dn22.png there is approximately 95.44% of the distribution.

3 . meanwhile, in the interval dn23.png it is comprised, approximately, 99.74% of the distribution. These properties are very useful for the establishment of confidence intervals. On the other hand, the fact that practically the entire distribution is at three standard deviations from the mean justifies the limits of the tables commonly used in the standard normal.

5 . yes dn24.png so dn25.png

6 . yes dn26.png are independent normal random variables, then:

  • Its sum is normally distributed with dn27.png (demonstration). Reciprocally, if two independent random variables have a normally distributed sum, they should be normal (Crámer's Theorem).

    • Your difference is normally distributed with dn28.png
  • If the variances of X and Y are equal, then U and V are independent of each other.

  • The divergence of Kullback-Leibler,
    dn29.png

7 . d38.png are independent random variables normally distributed, then:

  • Your product X Y follows a distribution with density p, given by

dn30.png
where Ko, is a modified Bessel function of the second type.

  • Its quotient follows a Cauchy distribution with dn31.png . In this way the Cauchy distribution is a special type of quotient distribution.

8 . Yes dn32.png are standard independent normal variables, then
dn33.png follows a χ² distribution with n degrees of freedom.

9 . Yes dn34.png are standard independent normal variables, then the sample mean dn35.png and the sample variance

dn36.pngThey are independent. This property characterizes normal distributions and helps explain why the F-test is not robust with respect to non-normality).

General formula of the normal distribution

d39.png

The Central Limit Theorem

The Central Limit Theorem states that under certain conditions (as they can be independent and identically distributed with finite variance), the sum of a large number of random variables is roughly distributed as a normal one.

The practical importance of the Central Limit Theorem is that the distribution function of the normal one can be used as an approximation of some other distribution functions. For example:

  • A binomial distribution of parameters n and p is approximately normal for large values of n, and p not too close to 0 or 1 (some books recommend using this approximation only if np and n (1 - p) are both at least 5; a continuity correction should be applied).
    The approximate normal has parameters μ = np, σ2 = np (1 - p).

  • A Poisson distribution with parameter λ is approximately normal for large values of λ.
    The approximate normal distribution has parameters μ = σ2 = λ.
    The accuracy of these approximations depends on the purpose for which they are needed and the rate of convergence to the normal distribution. The typical case is that such approximations are less precise in the tails of the distribution. The Berry-Esséen Theorem provides a general upper limit of the approximation error of the distribution function.

dn16.png

Graph of the distribution function of a normal with μ = 12 and σ = 3, approximating the distribution function of a binomial with n = 48 and p = 1/4.

Below are some examples of exercises where the basic properties of the normal distribution are applied, as well as the general formula.

IMG_20180108_101728.jpg

IMG_20180108_101817.jpg

IMG_20180108_101745.jpg

IMG_20180108_101754.jpg

References

1 . Basic statistics. ITM. 2007. Retrieved on December 12, 2017.
2 . Orrego, Juan José Manzano (2014). PROCUREMENT LOGISTICS. Ediciones Paraninfo, S.A.
3 . Gómez-Chacón, Inés Ma; Català, Claudi Alsina; Raig, Núria Planas; Rodríguez, Joaquim Giménez; Muñoz, Yuly Marsela Vanegas; Sirera, Marta Civil (October 4, 2010). Mathematics education and citizenship. Grao
Four . It is a consequence of the Central Limit Theorem
5 . Abraham de Moivre, "Approximatio ad Summam Terminorum Binomii (a + b) n in Seriem expansi" (printed on November 12, 1733 in London for a private edition). This pamphlet was reprinted in: (1) Richard C. Archibald (1926) "A rare pamphlet of Moivre and some of his discoveries," Isis, vol. 8, pages 671-683; (2) Helen M. Walker, "De Moivre on the law of normal probability" in David Eugene Smith, A Source Book in Mathematics [New York, New York: McGraw-Hill, 1929; reprint: New York, New York: Dover, 1959], vol. 2, pages 566-575; (3) Abraham De Moivre, The Doctrine of Chances (2nd ed.)
6 . Wussing, Hans (March 1998). «Lesson 10». Lessons in the History of Mathematics (1st (Spanish) edition). 21st Century of Spain Editores, S.A. p. 190. "" The normal distribution and its applications to the theory of errors is often associated with the name of Gauss, who discovered it - just as Laplace - independently, but had already been studied by de Moivre. "

Coin Marketplace

STEEM 0.20
TRX 0.12
JST 0.029
BTC 61182.87
ETH 3360.06
USDT 1.00
SBD 2.49