- •2 The classical definition of probability.
- •3)Geometric definition of probability. The problem of meeting. Geometric Probability
- •4)Elements of the combinatory
- •5) A permutation. The number of permutations of n objects.
- •Example
- •6)A combination. The number of combinations of n distinct objects taken r at a time.
- •An Example of Combinations
- •7. Additive Rules (Addition formula of probability). 1) a1, a2, . . . , An are mutually exclusive. 2) a1, a2, . . . , An eny events.
- •8. Conditional probability.
- •9. Independence
- •Independent events
- •11. The Theorem of total probability.
- •12. Bayes’ Rule.
- •13. Bernoulli scheme. Bernoulli distribution.
- •14. Poisson approximation formula.
- •15. The Local Moivre-Laplace’s theorem
- •18. Independence of random variables.
- •21. Discrete random variable
- •22. Discrete Probability Distributions. Probability Density function.
- •23. Discrete Probability Distributions
- •24. Continuous distribution function
- •25. Continuous distribution function
- •Example
- •27. Joint Density Function
- •28. Conditional distribution
- •29. Statistical Independence
- •Independent events
- •30. Mathematical expectation
- •31. Mathematical expectation to the case of two random variables
- •32. Variance of random variables
- •33. Standard deviation.
- •35. Covariance of Random Variables
- •36. The correlation coefficient.
- •37. Means and Variances of Linear Combinations of Random Variables. We are still working towards finding the theoretical mean and variance of the sample mean:
- •Example
- •38. Chebyshev’s Theorem.
- •Example
- •39. Some Discrete Probability Distributions. Binomial and Multinomial Distributions.
- •40. Geometric Distribution.
- •41. Poisson Distribution.
- •42. Continuous Uniform Distribution. Normal Distribution.
- •43. Exponential Distributions.
- •44. Moments and Moment-Generating Functions.
- •45. Populations and Samples. Some Important Statistics.
- •46. Location Measures of a Sample: The Sample Mean, Median, and Mode.
- •1.Sample Mean.
- •Note that the statistics X(отрицание) assumes the value
- •47. The Sample Variance, Standard Deviation and Range.
- •48. The Central Limit Theorem.
- •49. The Likelihood Function.
- •50. Point estimate.
- •51. Estimating the Mean.
- •53. Single Sample: Estimating the Variance.
- •54. Sampling Distribution of s2.
- •55. Statistical Hypotheses: General Concepts.
- •56. Prove the formula of Poisson distribution:
23. Discrete Probability Distributions
A discrete probability distribution shall be understood as a probability distribution characterized by a probability mass function. Thus, the distribution of a random variable X is discrete, and X is then called a discrete random variable, if
as u runs through the set of all possible values of X. It follows that such a random variable can assume only a finite or countably infinite number of values.
Def: Let X be a random variable that takes the numerical values X1, X2, ..., Xn with probablities p(X1), p(X2), ..., p(Xn) respectively. A discrete probability distribution consists of the values of the random variable X and their corresponding probabilities P(X). The probabilities P(X) are such that
∑ P(X) = 1
Cumulative distribution
Def : For every real number x, the cumulative distribution function of a real-valued random variable X is given by
where the right-hand side represents the probability that the random variable X takes on a value less than or equal to x. The probability that X lies in the interval (a, b], where a < b, is therefore
Here the notation (a, b], indicates a semi-closed interval.
If treating several random variables X, Y, ... etc. the corresponding letters are used as subscripts while, if treating only one, the subscript is omitted. It is conventional to use a capital F for a cumulative distribution function, in contrast to the lower-case f used for probability density functions and probability mass functions. This applies when discussing general distributions: some specific distributions have their own conventional notation, for example the normal distribution.
The CDF of a continuous random variable X can be defined in terms of its probability density function ƒ as follows:
F(x) has the following properties:
1. F(x) non-decreasing: F(x1)≤F(x2), for x1≤x2
2.
3. F(x) is continuous from the right Ɏh=R (x→+0) lim F(x+h)=F(x)
24. Continuous distribution function
Continuous distribution function-random variables that can take on values on a continuous scale
Def 1: the function f(x) is a probability density function for the continuous random variable x define over the set of real number
f(x) ≥0
P(a<x<b)=
Def 2: the cumulative distribution function F(x) of continuous random variable x with density function f(x)
F(x)=p(X≤x)=
where -∞<x<∞
Properties:
a<x<b P(a<x<b)=F(b)-F(b)
f(x) =
mean
µ=ᶋ x f(x)dx
standard deviation
where f(x)- is density function
σ2 = E((X-µ02))=ᶋ((x-µ) 2f(x)dx) where ᶋ -∞<x<∞
Probability density function
In probability theory, a probability density function, or density of a continuous random variable, is a function that describes the relative likelihood for this random variable to take on a given value. The probability for the random variable to fall within a particular region is given by the integral of this variable’s density over the region. The probability density function is nonnegative everywhere, and its integral over the entire space is equal to one.
A random variable X with values in a measure space (usually Rn with the Borel sets as measurable subsets) has as probability distribution the measure X∗P on : the density of X with respect to a reference measure μ on is the Radon–Nikodym derivative:
That is, f is any measurable function with the property that:
for any measurable set .
