[go: up one dir, main page]

0% found this document useful (0 votes)
4 views4 pages

Chapter 13 Probability

Chapter 13 covers key concepts in probability, including conditional probability, Bayes' theorem, and the definition of random variables. It explains the probability distribution of random variables, the calculation of mean and variance, and the characteristics of Bernoulli trials. The chapter also introduces the binomial distribution formula for calculating probabilities in a series of trials.

Uploaded by

kmakabe78
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
4 views4 pages

Chapter 13 Probability

Chapter 13 covers key concepts in probability, including conditional probability, Bayes' theorem, and the definition of random variables. It explains the probability distribution of random variables, the calculation of mean and variance, and the characteristics of Bernoulli trials. The chapter also introduces the binomial distribution formula for calculating probabilities in a series of trials.

Uploaded by

kmakabe78
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 4

Key Notes

Chapter-13

Probability

The salient features of the chapter are –

• The conditional probability of an event E, given the occurrence of the event F is given by

P(E ∩ F)
P ( E |F ) = , P(F) ≠ 0
P(F)

0 ≤ P ( E | F ) ≤ 1,
P ( E′ | F ) = 1 – P (E | F )
P (( E ∪ F ) G ) = P ( E G ) + P ( F | G ) – P (( E ∩ F ) | G )
P ( E ∩ F ) = P (E) P ( E | F ), P (E) ≠ 0

P (E ∩ F ) = P ( F ) P (E | F ), P (F ) ≠ 0

P (E ∩ F ) = P (E) P (F )
• P ( E | F ) = P (E), P (F ) ≠ 0
P ( F | E) = P (F ), P (E) ≠ 0

• Theorem of total probability:

Let {E1, E2 , ..., En ) be a partition of a sample space and suppose that each of E1 , E2 , ..., En
has non zero probability. Let A be any event associated with S, then

P(A) = P(E1 )P(A | E1 ) + P(E 2 ) + P(A | E 2 ) + ........ + P(E n )P(A | E n )

• Bayes' theorem: If E1 , E2 , ..., En are events which constitute a partition of sample space S,

i.e. E1 , E2 , ..., En are pairwise disjoint and E1 4, E2 4, ..., 4 En = S and A be any event with

P(Ei ) P(A | Ei )
non-zero probability, then, P ( Ei |A ) = n
∑ P(Ei ) P(A | Ei )
j1

• A random variable is a real valued function whose domain is the sample space of a random
experiment.

• The probability distribution of a random variable X is the system of numbers

Material Downloaded From SUPERCOP


Key Notes
X : x1 x2 ..... xn

P(X) : p1 p2 ...... pn

n
pi > o, ∑ pi = 1, i = 1, 2,......, n
Where, i =1

• Let X be a random variable whose possible values x1, x2, x3 .... ..., xn occur with probabilities
n
p1, p2, p3 .... ..., pn respectively. The mean of X, denoted by µ is the number ∑ x i pi . The mean
i =1

of a random variable X is also called the expectation of X, denoted by E (X).

• Let X be a random variable whose possible values x1, x2, x3 .... ..., xn occur with probabilities

p(x1 ), p(x 2 ),......p(x n ) respectively. Let µ = E ( X ) be the mean of X. The variance of X,


n
denoted by Var (X) or σ2x is defined as x 2 Var ( X ) = ∑ (x i µ) 2 p(xi ) or equivalently
i =1

n

2
σ2x = E ( X – µ ) . The non-negative number, x Va r ( X ) = (x i µ)2 p(x i ) is called the
1=i

standard deviation of the random variable X.

2
Var ( X ) = E X 2 ( ) –  E ( X ) 

• Trials of a random experiment are called Bernoulli trials, if they satisfy the following
conditions:

(i) There should be a finite number of trials.

(ii) The trials should be independent.

(iii) Each trial has exactly two outcomes: success or failure.

(iv) The probability of success remains the same in each trial.

For Binomial distribution B(n, p), P(X=x) =n Cx q n −x P x , x = 0, 1, ......, n(q = 1 − p)

Material Downloaded From SUPERCOP

You might also like