- •2 The classical definition of probability.
- •3)Geometric definition of probability. The problem of meeting. Geometric Probability
- •4)Elements of the combinatory
- •5) A permutation. The number of permutations of n objects.
- •Example
- •6)A combination. The number of combinations of n distinct objects taken r at a time.
- •An Example of Combinations
- •7. Additive Rules (Addition formula of probability). 1) a1, a2, . . . , An are mutually exclusive. 2) a1, a2, . . . , An eny events.
- •8. Conditional probability.
- •9. Independence
- •Independent events
- •11. The Theorem of total probability.
- •12. Bayes’ Rule.
- •13. Bernoulli scheme. Bernoulli distribution.
- •14. Poisson approximation formula.
- •15. The Local Moivre-Laplace’s theorem
- •18. Independence of random variables.
- •21. Discrete random variable
- •22. Discrete Probability Distributions. Probability Density function.
- •23. Discrete Probability Distributions
- •24. Continuous distribution function
- •25. Continuous distribution function
- •Example
- •27. Joint Density Function
- •28. Conditional distribution
- •29. Statistical Independence
- •Independent events
- •30. Mathematical expectation
- •31. Mathematical expectation to the case of two random variables
- •32. Variance of random variables
- •33. Standard deviation.
- •35. Covariance of Random Variables
- •36. The correlation coefficient.
- •37. Means and Variances of Linear Combinations of Random Variables. We are still working towards finding the theoretical mean and variance of the sample mean:
- •Example
- •38. Chebyshev’s Theorem.
- •Example
- •39. Some Discrete Probability Distributions. Binomial and Multinomial Distributions.
- •40. Geometric Distribution.
- •41. Poisson Distribution.
- •42. Continuous Uniform Distribution. Normal Distribution.
- •43. Exponential Distributions.
- •44. Moments and Moment-Generating Functions.
- •45. Populations and Samples. Some Important Statistics.
- •46. Location Measures of a Sample: The Sample Mean, Median, and Mode.
- •1.Sample Mean.
- •Note that the statistics X(отрицание) assumes the value
- •47. The Sample Variance, Standard Deviation and Range.
- •48. The Central Limit Theorem.
- •49. The Likelihood Function.
- •50. Point estimate.
- •51. Estimating the Mean.
- •53. Single Sample: Estimating the Variance.
- •54. Sampling Distribution of s2.
- •55. Statistical Hypotheses: General Concepts.
- •56. Prove the formula of Poisson distribution:
Independent events
The standard definition says:
Here A ∩ B is the intersection of A and B, that is, it is the event that both events A and B occur.
More generally, any collection of events—possibly more than just two of them—are mutually independent if and only if for every finite subset A1, ..., An of the collection we have
This is called the multiplication rule for independent events. Notice that independence requires this rule to hold for everysubset of the collection; see[2] for a three-event example in which and yet no two of the three events are pairwise independent.
If two events A and B are independent, then the conditional probability of A given B is the same as the unconditional (or marginal) probability of A, that is,
There are at least two reasons why this statement is not taken to be the definition of independence: (1) the two events A and B do not play symmetrical roles in this statement, and (2) problems arise with this statement when events of probability 0 are involved.
The conditional probability of event A given B is given by
The statement above, when is equivalent to
which is the standard definition given above.
Note that an event is independent of itself if and only if
That is, if its probability is one or zero. Thus if an event or itscomplement almost surely occurs, it is independent of itself. For example, if event A is choosing any number but 0.5 from auniform distribution on the unit interval, A is independent of itself, even though, tautologically, A fully determines A.
30. Mathematical expectation
Example: we have 3 red balls and 4 blue balls in a box. At randomly getting we choose 2 balls. Find the probability that 2 balls are red.
Sample Space |
X |
RR |
2 |
RB |
1 |
BR |
1 |
BB |
0 |
X-random variable
P(X)= 2* 2/7 + 1* 2/7 + 0*0= 6/7
Def.1. A random variable is a function what associated a real number with each element is a sample.
Def.2. The set of ordered pairs (x, f(x)) is a probability function or probability mass function or probability distribution of a discrete random variable, if for each possible outcomes x is equal
f(x) ≥0
∑(x=0, to ∞ )=1
P(X=x)=f(x)
Def.3. Let X is a random variable with probability distribution f(x). The mean or expected value of x
µ=E(X)= ∑x f(x)
Def 4 Let x be a random variable with probability distribution f (x) and expected value µ the variance of x is equal
σ 2 =E(x-µ)2= ∑ (x-µ)f(x)
def 5 Let x random variable with probability f(x) and expected value µ the variance of x= σ 2
σ 2 =E(x-µ)2=∑(x-µ)2 f(x)
def 6 let x and y random variable with joint probability distribution f(x,y) the covariance of x,y we denote
σ x,y= E(x-µx)(y-µy)=∑(x-µx)(y-µy) f(x,y)
def 7 let x,y be a random variable with covariance σ x,y and standart deviation σx and σy respectively. The correlation:
ƥ
xy=
Theorem 2 the variance of a random variable x is
σ2=E (x 2)-µ2
