STATISTICS
STATISTICS
STATISTICS
KNOWLEDGE ARENA
(BEST COACHING FOR UGC NTA NET/JRF)
Statistics Notes
Points to be remembered
F distribution is coined by George W Snedewr in Honour of Sir Ronald A Fisher.
Chi square (non- parametric test) concept given by Karl Pearson.
Concept of normal distribution is given by De Mouire and person involved in this are Laplace, Gauss and W J
Yoden.
Concept of Regression is given by Sir Francis Galton in 1877.
Concept of T- distribution is given by WS Gooset.
Data which are collected for the first time are primary data.
Data is only quantitative.
Secondary data are second hand data which are in the form of published & unpublished.
Category of secondary data are also called paper source.
Median and mode are positional average.
Athematic mean, geometric, harmonic & weighted average mean are mathematical average.
Σ known as capital SIGMA.
Property of arithmetic mean-:
1) Sum of all observations of the given set of observations from their arithmetic mean is zero.
3) The sum of square of the deviations of the given set of observations is minimum when taken from the arithmetic mean.
5) AM> GM> HM ( AM- arithmetic mean, GM- geometric mean, HM- harmonic mean)
One dimensional 1d diagram are those which have only length. Examples are line diagram, multiple bar
diagram, compound or cluster bar diagram, sub divided bar diagram are also called Component Bar Diagram,
Percentage Bar Diagram, Deviational Bar Diagram.
Two dimensional diagram are those in which both length and breadth are present 2D. Examples are
histogram, area diagrams, rectangles, square, circles and pie diagram.
OGIVE CURVE and frequency polygon are also 2d diagrams.
OGIVE represent cumulative frequency and histogram frequency distribution and frequency polygon
The larger the sample the more accurate will be the research.
Increasing the sample size decreases the sample error.
www.knowledgearena.in Page 1
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
Type 1 error by rejecting a true null hypothesis and it is also known as producer error, alpha or level of
significance.
Type 2 error by accepting the false null hypothesis also called consumer error (1- β) beta, power function of
test or power curve, power of test.
Standard error (SE) is standard deviation of the distribution of the sample mean. S.E = σ/√Ŷ
In a normal distribution curve, since the curve is a bell shaped & symmetrical i.e. mean=median=mode
Total area under normal probability curve is 1 (.5 + .5)
Since curve is symmetrical co efficient of kurtosis is 3 mesocratic.
Range of distribution is ∞to ∞ but practically it is 60σ.
Point of inflexion is x= +- μσ
Leptokurtic <3, Platokurtic>, Mesocratic =3.
Area μ+-1σ= 68.27%, μ+-2σ= 95.45%, μ+- 3σ= 99.73%
Z (STANDARD NORMAL DISTRIBUTION) = X-μ/Σ
Concept of BINOMIAL DISTRIBUTION is given by JAMES BERNOULLI.
Concept of POISSON DISTRIBUTION is given by SIMEON POISSON. In this the value of mean and
variance is (0,0)
Standard deviation is also known as root Mean Square Deviation.
Standard deviation is affected by change of scale & independent of change of origin.
Positively skewed= mean>median>mode.
Negatively skewed= mode>median>mean.
Balance pattern= mean=median=mode.
Skewness means lack of symmetry or asymmetrical distribution.
Concept of co efficient of Skewness given by Karl Pearson.
Confidence interval 95%= 1.96, 99%= 2.56/2.58
Chi square is a non- parametric test
Chi square lies from 0 to ∞
Conditions to apply chi square test:-
1) Population mean & sample mean is not given in the question.
2) Degree of freedom (df-1) as df starts from t test.
Chi square test is also known as:-
1) Goodness of fit accumulation.
2) Contingency table.
3) Quantitative variables.
4) Co efficient of association.
Degree of freedom = (row-1) (column -1)
www.knowledgearena.in Page 2
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
The most common basic statistics terms you’ll come across are the mean, mode and median. These are all what
are known as “Measures of Central Tendency.” Also important in this early chapter of statistics is the shape of a
distribution. This tells us something about how data is spread out around the mean or median. Perhaps the most
common distribution you’ll see is the normal distribution, sometimes called a bell curve. Heights, weights, and
many other things found in nature tend to be shaped like this:
www.knowledgearena.in Page 3
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
Overview
Stuck on how to find the mean, median, & mode in statistics?
“A la mode” is a French word that means fashionable and it also refers to a popular way of serving ice
cream. So “Mode” is the most popular or fashionable member of a set of numbers. The word MOde is also
like MOst.
The “Mean” requires you do arithmetic (adding all the numbers and dividing) so that’s the “mean” one.
“Median” has the same number of letters as “Middle”.
The Mean
Mean vs. Median
Mean vs. Average
Specific “Means” commonly used in Stats
Other Types
Mean vs Median
www.knowledgearena.in Page 4
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
Both are measures of where the center of a data set lies, but they are usually different numbers. For example, take this list
of numbers: 10,10,20,40,70.
The mean (average) is found by adding all of the numbers together and dividing by the number of items in the set: 10
+ 10 + 20 + 40 + 70 / 5 = 30.
The median is found by ordering the set from lowest to highest and finding the exact middle. The median is just the
middle number: 20.
Sometimes the two will be the same number. For example, the data set 1,2,4,6,7 has an average of 1 + 2 + 4 + 6 + 7 / 5 = 4
and a median (a middle) of 4.
Mean of the sampling distribution: used with probability distributions, especially with the Central Limit Theorem.
It’s an average of a set of distributions.
Sample mean: the average value in a sample.
Population mean: the average value in a population.
Other Types
There are other types of means, and you’ll use them in various branches of math. Most have very narrow applications to
fields like finance or physics; if you’re in elementary statistics you probably won’t work with them.
These are some of the most common types you’ll come across.
1. Weighted mean.
2. Harmonic mean.
3. Geometric mean.
4. Arithmetic-Geometric mean.
5. Root-Mean Square mean.
6. Heronian mean.
1. Weighted Mean
These are fairly common in statistics, especially when studying populations. Instead of each data point contributing
equally to the final average, some data points contribute more than others. If all the weights are equal, then this will
equal the arithmetic mean. There are certain circumstances when this can give incorrect information, as shown
by Simpson’s Paradox.
www.knowledgearena.in Page 5
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
2. Harmonic Mean
To find it:
A. Add the reciprocals of the numbers in the set. To find a reciprocal, flip the fraction so that the numerator
becomes the denominator and the denominator becomes the numerator. For example, the reciprocal of 6/1 is
1/6.
B. Divide the answer by the number of items in the set.
C. Take the reciprocal of the result.
This is used quite a lot in physics. In some cases involving rates and ratios it gives a better average than the
arithmetic mean. You’ll also find uses in geometry, finance and computer science.
3. Geometric Mean
4. Arithmetic-Geometric Mean
This is used mostly in calculus and in machine computation (i.e. as the basic for many computer calculations). It’s
related to the perimeter of an ellipse. When it was first developed by Gauss, it was used to calculate planetary orbits.
The arithmetic-geometric is (not surprisingly!) a blend of the arithmetic and geometric averages. The math is quite
complicated but you can find a relatively simple explanation of the math here.
www.knowledgearena.in Page 6
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
Sampling error is the tendency for a statistic not to exactly match the population. Error doesn’t necessarily mean that a
mistake was made in your sampling; Sampling Variability could be a more accurate name. For example, let’s say you
have a population in the United States with an average height of 5 feet 9 inches. If you take a sample, even a fairly sizable
sample of say, 10,000 people, it’s unlikely that you’ll get exactly 5 feet 9 inches. You might get very close, perhaps to
within a fraction of an inch. If you repeat the experiment, you might get another very close result. For example, in
experiment 1 you might get 5 feet 8.9 inches and in experiment 2 you might get 5 feet 9.1 inches. The tendency for
statistics to get very close, but not exactly right, is called sampling error.
Note: If the statistic is unbiased, the average of all statistics from all samples will average the true population parameter.
Measurement Errors
Measurement errors are where a provided response is different from the real value. For example, you might survey to
find out if a person voted for President Obama. A person may have voted for him, but they are confused by the wording of
the questionnaire and mistakenly respond that they did not vote for him. Several factors may cause measurement error,
including:
The way the interviewer poses the question.
The wording on the questionnaire.
The way the data is collected.
The respondent’s record-keeping system.
Biased Estimators
In statistics, an estimator is a rule for calculating an estimate of a quantity based on observed data. For example, you
might have a rule to calculate a population mean. The result of using the rule is an estimate (a statistic) that hopefully is a
true reflection of the population. The bias of an estimator is the difference between the statistic’s expected value and the
true value of the population parameter. If the statistic is a true reflection of a population parameter it is
an unbiased estimator. If it is not a true reflection of a population parameter it is a biased estimator.
www.knowledgearena.in Page 7
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
The word bias in the regular English language implies that you have a personal reason to misrepresent a piece of
information. However, in statistics, it doesn’t mean that the interviewer, the researcher or even the respondent in an
interview is biased in some way. It just means that the estimator being used doesn’t produce a good estimate.
Breeds of dog are qualitative variables. How many dogs are quantitative variables.
Decimals Cities
www.knowledgearena.in Page 8
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
Exponents Tribe
As a general rule, if you can apply some kind of math (like addition), it’s a quantitative variable. Otherwise, it’s
qualitative. For example, you can’t add blue+green (unless you’re in an art class — even then you “mix” them, you don’t
add them!).
Numbers are sometimes assigned to qualitative variables for data analysis, but they are still classified as qualitative
variables despite the numerical classification. For example, a study may assign the number “1” to males and “2” to
females.
www.knowledgearena.in Page 9
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
You could turn age into a discrete variable and then you could count it. For example:
A person’s age in years.
A baby’s age in months.
Take a look at this article on orders of magnitude of time and you’ll see why time or age just isn’t countable. Try counting
your age in Planctoseconds (good luck…see you at the end of time!).
www.knowledgearena.in Page 10
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
Not paying your power bill on time and having your power cut off.
Boarding a plane first and finding a good seat.
Parking illegally and getting a parking ticket. Parking illegally increases your odds of getting a ticket.
Buying ten lottery tickets and winning the lottery. The more tickets you buy, the greater your odds of winning.
Driving a car and getting in a traffic accident.
In math, a parameter is something in an equation that is passed on in an equation. It means something different in
statistics. It’s a value that tells you something about a population and is the opposite from a statistic, which tells you
something about a small part of the population.
www.knowledgearena.in Page 11
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
A parameter never changes, because everyone (or everything) was surveyed to find the parameter. For example, you
might be interested in the average age of everyone in your class. Maybe you asked everyone and found the average age
was 25. That’s a parameter, because you asked everyone in the class. Now let’s say you wanted to know the average age
of everyone in your grade or year. If you use that information from your class to take a guess at the average age, then that
information becomes a statistic. That’s because you can’t be sure your guess is correct (although it will probably be
close!).
Statistics vary. You know the average age of your classmates is 25. You might guess that the average age of everyone in
your year is 24, 25, or 26. You might guess the average age for other colleges in your area is the same. And you might
even guess that’s the average age for college students in the U.S.. These may not be bad guesses, but they are statistics
because you didn’t ask everyone.
Body types are varied — they come in all shapes and sizes.
www.knowledgearena.in Page 12
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
Standard deviation is a measure of dispersement in statistics. “Dispersement” tells you how much your data is spread out.
Specifically, it shows you how much your data is spread out around the mean or average. For example, are all your scores
close to the average? Or are lots of scores way above (or way below) the average score?
The following graph of a normal distribution represents a great deal of data in real life. The mean, or average, is
represented by the Greek letter μ, in the center. Each segment (colored in dark blue to light blue) represents one standard
deviation away from the mean. For example, 2σ means two standard deviations from the mean.
Statistics Symbols A to Z
www.knowledgearena.in Page 13
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
Ho = null hypothesis.
H1 or Ha: alternative hypothesis.
IQR = interquartile range.
m = slope of a line.
M: median.
n: sample size or number of trials in a binomial experiment.
N: population size.
σ: standard deviation.
σx̅: standard error of the mean.
σp̂: standard error of the proportion.
p: p-value, or probability of success in a binomial experiment, or population proportion.
ρ: correlation coefficient for a population.
p̂: sample proportion.
P(A): probability of event A.
P(AC) or P(not A): the probability that A doesn’t happen.
P(B|A): the probability that event B occurs, given that event A occurs.
q: probability of failure in a binomial or geometric distribution.
Q1: first quartile.
Q3: third quartile.
r: correlation coefficient of a sample.
R2: coefficient of determination.
s: standard deviation of a sample.
s.d or SD: standard deviation.
t: t-score.
μ mean.
ν: degrees of freedom.
X: a variable.
Χ2: chi-square.
x: one data value.
x̄: mean of a sample.
z: z-score.
www.knowledgearena.in Page 14
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
– x
The probability density function (pdf) for this distribution is px (1 – p)1 , which can also be written as:
The expected value for a random variable, X, from a Bernoulli distribution is:
E[X]=p.
www.knowledgearena.in Page 15
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
Coin tossing as a game of probability and chance has been around since Roman times.
Bernoulli trials are usually phrased in terms of success and failure. Success doesn’t mean success in the usual way — it
just refers to an outcome you want to keep track of. For example, you might want to find out how many boys are born
each day, so you call a boy birth a “success” and a girl birth a “failure.” In the dice rolling example, a double six die roll
would be your “success” and everything else rolled would be considered a “failure.”
Independence
An important part of every Bernoulli trial is that each action must be independent. That means the probabilities must
remain the same throughout the trials; each event must be completely separate and have nothing to do with the previous
event.
Winning a scratch off lottery is an independent event. Your odds of winning on one ticket are the same as winning on any
other ticket. On the other hand, drawing lotto numbers is a dependent event. Lotto numbers come out of a ball (the
numbers aren’t replaced) so the probability of successive numbers being picked depends upon how many balls are left;
when there’s a hundred balls, the probability is 1/100 that any number will be picked, but when there are only ten balls
left, the probability shoots up to 1/10. While it’s possible to find those probabilities, it isn’t a Bernoulli trial because the
events (picking the numbers) are connected to each other.
The Bernouilli process leads to several probability distributions:
The binomial distribution,
The geometric distribution,
The negative binomial distribution.
www.knowledgearena.in Page 16
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
The Bernoulli distribution is closely related to the Binomial distribution. As long as each individual Bernoulli trial is
independent, then the number of successes in a series of Bernoulli trails has a Binomial Distribution. The Bernoulli
distribution can also be defined as the Binomial distribution with n = 1.
Criteria
Binomial distributions must also meet the following three criteria:
1. The number of observations or trials is fixed. In other words, you can only figure out the probability of
something happening if you do it a certain number of times. This is common sense—if you toss a coin once, your
probability of getting a tails is 50%. If you toss a coin a 20 times, your probability of getting a tails is very, very
close to 100%.
2. Each observation or trial is independent. In other words, none of your trials have an effect on the probability of
the next trial.
3. The probability of success (tails, heads, fail or pass) is exactly the same from one trial to another.
A binomial distribution is a simple experiment where there is “success” or “failure.” For example, choosing a winning
lottery ticket could be a binomial experiment (you either win or lose!). Tossing a coin to try and get heads is also binomial
(with tossing a heads being a “success” and a tails a “failure”). The formula for the variance of binomial distribution
is n*p (1-p) or n*p*q. The two formulas are equivalent because q = (1-p).
Example problem: If you flip a coin 50 times and try to get heads, what is the variance of binomial distribution?
Step 1: Find “p”. The first step to solving this problem is to realize that the probability of getting a heads is 50 percent, or
.5. Therefore, “p” (the probability) is .5.
Step 2: Find “q”, or 1-p. These two are equivalent. They are the probability of not getting a heads (in other words, the
probability of getting a tails). 1 – 0.5 = 0.5. Therefore, “q” (or 1 – p) = 0.5.
Step 3: Multiply Step 1 (p) by Step 2 (q) by “n” (the number of trials). We are flipping the coin 50 times, so the number of
trials is 50 (n = 50).
N * p * q = 50 * .5 * .5 = 12.5.
The var. of binomial distribution for flipping a coin 50 times is 12.5.
In essence, not a lot! The variance isn’t used for much at all, except for calculating standard deviation. For example, the
standard deviation for this particular binomial distribution is:
√12.5=3.54.
You’ll use the variance for things like calculating z-scores (this typically comes later in a stats class, after normal
distributions), which has a standard deviation in the bottom of the formula:
www.knowledgearena.in Page 17
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
Population Variance
The population variance is a type of parameter. If you aren’t sure what a parameter is, you may want to review:
What is the Difference Between a Statistic and a Parameter?
The formula is:
Formula:
Example problem: Find standard deviation for a binomial distribution with n = 5 and p = 0.12.
Step 1: Subtract p from 1 to find q.
1 – .12
=.88
Step 2: Multiply n times p times q.
5 * .12 * .88
=.528
Step 3: Find the square root of the answer from Step 2.
√.528 = =.727 (rounded to 3 decimal places).
Many groups follow this type of pattern. That’s why it’s widely used in business, statistics and in government bodies like
the FDA:
Heights of people.
Measurement errors.
www.knowledgearena.in Page 18
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
Blood pressure.
Points on a test.
IQ scores.
Salaries.
The empirical rule tells you what percentage of your data falls within a certain number of standard deviations from
the mean:
• 68% of the data falls within one standard deviation of the mean.
• 95% of the data falls within two standard deviations of the mean.
• 99.7% of the data falls within three standard deviations of the mean.
The standard deviation controls the spread of the distribution. A smaller standard deviation indicates that the data is
tightly clustered around the mean; the normal distribution will be taller. A larger standard deviation indicates that the data
is spread out around the mean; the normal distribution will be flatter and wider.
www.knowledgearena.in Page 19
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
The standard normal distribution could help you figure out which subject you are getting good grades in and which
subjects you have to exert more effort into due to low scoring percentages. Once you get a score in one subject that is
higher than your score in another subject, you might think that you are better in the subject where you got the higher
score. This is not always true.
You can only say that you are better in a particular subject if you get a score with a certain number of standard deviations
above the mean. The standard deviation tells you how tightly your data is clustered around the mean; It allows you to
compare different distributions that have different types of data — including different means.
For example, if you get a score of 90 in Math and 95 in English, you might think that you are better in English than in
Math. However, in Math, your score is 2 standard deviations above the mean. In English, it’s only one standard deviation
above the mean. It tells you that in Math, your score is far higher than most of the students (your score falls into the tail).
Based on this data, you actually performed better in Math than in English!
www.knowledgearena.in Page 20
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
Poisson distributions, valid only for integers on the horizontal axis. λ (also written as μ) is the expected number of event
occurrences.
Practical Uses of the Poisson Distribution:
A textbook store rents an average of 200 books every Saturday night. Using this data, you can predict the probability
that more books will sell (perhaps 300 or 400) on the following Saturday nights. Another example is the number of
diners in a certain restaurant every day. If the average number of diners for seven days is 500, you can predict the
probability of a certain day having more customers.
Because of this application, Poisson distributions are used by businessmen to make forecasts about the number of
customers or sales on certain days or seasons of the year. In business, overstocking will sometimes mean losses if the
goods are not sold. Likewise, having too few stocks would still mean a lost business opportunity because you were not
able to maximize your sales due to a shortage of stock. By using this tool, businessmen are able to estimate the time when
demand is unusually higher, so they can purchase more stock. Hotels and restaurants could prepare for an influx of
customers, they could hire extra temporary workers in advance, purchase more supplies, or make contingency plans just in
case they cannot accommodate their guests coming to the area.
With the Poisson distribution, companies can adjust supply to demand in order to keep their business earning good profit.
In addition, waste of resources is prevented.
Calculating the Poisson Distribution
The Poisson Distribution pmf is: P(x; μ) = (e-μ * μx) / x!
Where:
The symbol “!” is a factorial.
μ (the expected number of occurrences) is sometimes written as λ. Sometimes called the event rate or rate
parameter.
Example question
The average number of major storms in your city is 2 per year. What is the probability that exactly 3 storms will
hit your city next year?
Step 1: Figure out the components you need to put into the equation.
μ = 2 (average number of storms per year, historically)
x = 3 (the number of storms we think might hit next year)
e = 2.71828 (e is Euler’s number, a constant)
Step 2: Plug the values from Step 1 into the Poisson distribution formula:
www.knowledgearena.in Page 21
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
Hypothesis Testing
What is a Hypothesis?
If you are going to propose a hypothesis, it’s customary to write a statement. Your statement will look like this:
“If I…(do this to an independent variable)….then (this will happen to the dependent variable).”
For example:
If I (decrease the amount of water given to herbs) then (the herbs will increase in size).
If I (give patients counseling in addition to medication) then (their overall depression scale will decrease).
If I (give exams at noon instead of 7) then (student test scores will improve).
If I (look in this certain location) then (I am more likely to find new species).
www.knowledgearena.in Page 22
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
Hypothesis testing in statistics is a way for you to test the results of a survey or experiment to see if you have meaningful
results. You’re basically testing whether your results are valid by figuring out the odds that your results have happened by
chance. If your results may have happened by chance, the experiment won’t be repeatable and so has little use.
Hypothesis testing can be one of the most confusing aspects for students, mostly because before you can even perform a
test, you have to know what your null hypothesis is. Often, those tricky word problems that you are faced with can be
difficult to decipher. But it’s easier than you think; all you need to do is:
The hypothesis statement in this question is that the researcher believes the average recovery time is more than 8.2 weeks.
It can be written in mathematical terms as:
H1: μ > 8.2
Next, you’ll need to state the null hypothesis (See: How to state the null hypothesis). That’s what will happen if the
researcher is wrong. In the above example, if the researcher is wrong then the recovery time is less than or equal to 8.2
weeks. In math, that’s:
H0 μ ≤ 8.2
Rejecting the null hypothesis
Ten or so years ago, we believed that there were 9 planets in the solar system. Pluto was demoted as a planet in 2006. The
null hypothesis of “Pluto is a planet” was replaced by “Pluto is not a planet.” Of course, rejecting the null hypothesis isn’t
always that easy — the hard part is usually figuring out what your null hypothesis is in the first place.
www.knowledgearena.in Page 23
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
A principal at a certain school claims that the students in his school are above average intelligence. A random sample of
thirty students IQ scores have a mean score of 112. Is there sufficient evidence to support the principal’s claim? The mean
population IQ is 100 with a standard deviation of 15.
Step 1: State the Null hypothesis. The accepted fact is that the population mean is 100, so: H0: μ=100.
Step 2: State the Alternate Hypothesis. The claim is that the students have above average IQ scores, so:
H1: μ > 100.
The fact that we are looking for scores “greater than” a certain point means that this is a one-tailed test.
Step 4: State the alpha level. If you aren’t given an alpha level, use 5% (0.05).
Step 5: Find the rejection region area (given by your alpha level above) from the z-table. An area of .05 is equal to a z-
score of 1.645.
www.knowledgearena.in Page 24
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
Step 7: If Step 6 is greater than Step 5, reject the null hypothesis. If it’s less than Step 5, you cannot reject the null
hypothesis. In this case, it is greater (4.56 > 1.645), so you can reject the null.
What is a Z Test?
A Z-test is a type of hypothesis test. Hypothesis testing is just a way for you to figure out if results from a test are valid or
repeatable. For example, if someone said they had found a new drug that cures cancer, you would want to be sure it was
probably true. A hypothesis test will tell you if it’s probably true, or probably not true. A Z test, is used when your data is
approximately normally distributed.
What is an F Test?
www.knowledgearena.in Page 25
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
An “F Test” is a catch-all term for any test that uses the F-distribution. In most cases, when people talk about the F-Test,
what they are actually talking about is The F-Test to Compare Two Variances. However, the f-statistic is used in a variety
of tests including regression analysis, the Chow test and the Scheffe Test (a post-hoc ANOVA test).
Assumptions:
Several assumptions are made for the test. Your population must be approximately normally distributed (i.e. fit the
shape of a bell curve) in order to use the test. Plus, the samples must be independent events. In addition, you’ll want to
bear in mind a few important points:
The larger variance should always go in the numerator (the top number) to force the test into a right-tailed test.
Right-tailed tests are easier to calculate.
For two-tailed tests, divide alpha by 2 before finding the right critical value.
If you are given standard deviations, they must be squared to get the variances.
If your degrees of freedom aren’t listed in the F Table, use the larger critical value. This helps to avoid the
possibility of Type I errors.
The difference between running a one or two tailed F test is that the alpha level needs to be halved for two tailed F tests.
For example, instead of working at α = 0.05, you use α = 0.025; Instead of working at α = 0.01, you use α = 0.005.
With a two tailed F test, you just want to know if the variances are not equal to each other. In notation:
Ha = σ21 ≠ σ2 2
www.knowledgearena.in Page 26
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
Step 2: Calculate your F critical value. Put the highest variance as the numerator and the lowest variance as the
denominator:
F Statistic = variance 1/ variance 2 = 109.63 / 65.99 = 1.66
Step 4: Choose an alpha level. No alpha was stated in the question, so use 0.05 (the standard “go to” in statistics). This
needs to be halved for the two-tailed test, so use 0.025.
Step 5: Find the critical F Value using the F Table. There are several tables, so make sure you look in the alpha = .025
table. Critical F (40,20) at alpha (0.025) = 2.287.
Step 6: Compare your calculated value (Step 2) to your table value (Step 5). If your calculated value is higher than the
table value, you can reject the null hypothesis:
F calculated value: 1.66
F value from table: 2.287.
1.66 < 2 .287.
So we cannot reject the null hypothesis
What is a T test?
www.knowledgearena.in Page 27
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
The t test tells you how significant the differences between groups are; In other words it lets you know if those differences
(measured in means/averages) could have happened by chance.
A very simple example: Let’s say you have a cold and you try a naturopathic remedy. Your cold lasts a couple of days.
The next time you have a cold, you buy an over-the-counter pharmaceutical and the cold lasts a week. You survey your
friends and they all tell you that their colds were of a shorter duration (an average of 3 days) when they took the
homeopathic remedy. What you really want to know is, are these results repeatable? A t test can tell you by comparing the
means of the two groups and letting you know the probability of those results happening by chance
T-Test Assumptions
1. The first assumption made regarding t-tests concerns the scale of measurement. The assumption for a t-test is that
the scale of measurement applied to the data collected follows a continuous or ordinal scale, such as the scores for
an IQ test.
2. The second assumption made is that of a simple random sample, that the data is collected from a representative,
randomly selected portion of the total population.
3. The third assumption is the data, when plotted, results in a normal distribution, bell-shaped distribution curve.
When a normal distribution is assumed, one can specify a level of probability (alpha level, level of
significance, p) as a criterion for acceptance. In most cases, a 5% value can be assumed.
4. The fourth assumption is a reasonably large sample size is used. A larger sample size means the distribution of
results should approach a normal bell-shaped curve.
5. The final assumption is homogeneity of variance. Homogeneous, or equal, variance exists when the standard
deviations of samples are approximately equal.
www.knowledgearena.in Page 28
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
Uses
The T Distribution (and the associated t scores), are used in hypothesis testing when you want to figure out if you
should accept or reject the null hypothesis.
The central region on this graph is the acceptance area and the tail is the rejection region, or regions. In this particular
graph of a two tailed test, the rejection region is shaded blue. The area in the tail can be described with z-scores or t-
scores. For example, the image to the left shows an area in the tails of 5% (2.5% each side). The z-score would be 1.96
(from the z-table), which represents 1.96 standard deviations from the mean. The null hypothesis will be rejected if z is
less than -1.96 or greater than 1.96.
In general, this distribution is used when you have a small sample size (under 30) or you don’t know the population
standard deviation. For practical purposes (i.e. in the real world), this is nearly always the case. So, unlike in
your elementary statistics class, you’ll likely be using it in real life situations more than the normal distribution. If the size
of your sample is large enough, the two distributions are practically the same.
The T Score.
The t score is a ratio between the difference between two groups and the difference within the groups. The larger the t
score, the more difference there is between groups. The smaller the t score, the more similarity there is between groups. A
t score of 3 means that the groups are three times as different from each other as they are within each other. When you run
a t test, the bigger the t-value, the more likely it is that the results are repeatable.
A large t-score tells you that the groups are different.
A small t-score tells you that the groups are similar
www.knowledgearena.in Page 29
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
With a “regular” two sample t test, you’re comparing the means for two different samples. For example, you might test
two different groups of customer service associates on a business-related test or testing students from two universities on
their English skills. If you take a random sample each group separately and they have different conditions, your samples
are independent and you should run an independent samples t test (also called between-samples and unpaired-samples).
The null hypothesis for the for the independent samples t-test is μ1 = μ2. In other words, it assumes the means are equal.
With the paired t test, the null hypothesis is that the pairwise difference between the two tests is equal (H0: µd = 0). The
difference between the two tests is very subtle; which one you choose is based on your data collection method.
Paired Samples T Test By hand
Sample question: Calculate a paired t test by hand for the following data:
www.knowledgearena.in Page 30
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
www.knowledgearena.in Page 31
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
Step 6: Subtract 1 from the sample size to get the degrees of freedom. We have 11 items, so 11-1 = 10.
Step 7: Find the p-value in the t-table, using the degrees of freedom in Step 6. If you don’t have a specified alpha level,
use 0.05 (5%). For this sample problem, with df=10, the t-value is 2.228.
Step 8: Compare your t-table value from Step 7 (2.228) to your calculated t-value (-2.74). The calculated t-value is greater
than the table value at an alpha level of .05. The p-value is less than the alpha level: p <.05. We can reject the null
hypothesis that there is no difference between means.
Note: You can ignore the minus sign when comparing the two t-values, as ± indicates the direction; the p-value remains
the same for both directions.
www.knowledgearena.in Page 32
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
A chi-square goodness of fit test determines if a sample data matches a population. For more details on this type,
see: Goodness of Fit Test.
A chi-square test for independence compares two variables in a contingency table to see if they are related. In a
more general sense, it tests to see whether distributions of categorical variables differ from each another.
A very small chi square test statistic means that your observed data fits your expected data extremely
well. In other words, there is a relationship.
A very large chi square test statistic means that the data does not fit very well. In other words, there
isn’t a relationship.
The formula for the chi-square statistic used in the chi square test is:
Uses
The chi-squared distribution has many uses in statistics, including:
Confidence interval estimation for a population standard deviation of a normal distribution from a sample
standard deviation.
Independence of two criteria of classification of qualitative variables.
Relationships between categorical variables (contingency tables).
Sample variance study when the underlying distribution is normal.
Tests of deviations of differences between expected and observed frequencies (one-way tables).
The chi-square test (a goodness of fit test).
An ANOVA test is a way to find out if survey or experiment results are significant. In other words, they help you to figure
out if you need to reject the null hypothesis or accept the alternate hypothesis.
Basically, you’re testing groups to see if there’s a difference between them. Examples of when you might want to test
different groups:
A group of psychiatric patients are trying three different therapies: counseling, medication and biofeedback. You
want to see if one therapy is better than the others.
A manufacturer has two different processes to make light bulbs. They want to know if one process is better than
the other.
Students from different colleges take the same exam. You want to see if one college outperforms the other.
www.knowledgearena.in Page 33
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
Let’s say you are studying if an alcoholic support group and individual counseling combined is the most effective
treatment for lowering alcohol consumption. You might split the study participants into three groups or levels:
Medication only,
Medication and counseling,
Counseling only.
Your dependent variable would be the number of alcoholic beverages consumed per day.
If your groups or levels have a hierarchical structure (each level has unique subgroups), then use a nested ANOVA for the
analysis.
Types of Tests.
There are two main types: one-way and two-way. Two-way tests can be with or without replication.
One-way ANOVA between groups: used when you want to test two groups to see if there’s a difference between
them.
Two way ANOVA without replication: used when you have one group and you’re double-testing that same
group. For example, you’re testing one set of individuals before and after they take a medication to see if it works
or not.
Two way ANOVA with replication: Two groups, and the members of those groups are doing more than one
thing. For example, two groups of patients from different hospitals trying two different therapies.
www.knowledgearena.in Page 34
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
A one way ANOVA will tell you that at least two groups were different from each other. But it won’t tell you which
groups were different. If your test returns a significant f-statistic, you may need to run an ad hoc test (like the Least
Significant Difference test) to tell you exactly which groups had a difference in means.
The factors can be split into levels. In the above example, income level could be split into three levels: low, middle and
high income. Gender could be split into three levels: male, female, and transgender. Treatment groups are all possible
combinations of the factors. In this example there would be 3 x 3 = 9 treatment groups.
Two null hypotheses are tested if you are placing one observation in each cell. For this example, those hypotheses would
be:
H01: All the income groups have equal mean stress.
H02: All the gender groups have equal mean stress.
For multiple observations in cells, you would also be testing a third hypothesis:
H03: The factors are independent or the interaction effect does not exist.
An F-statistic is computed for each hypothesis you are testing.
www.knowledgearena.in Page 35
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
Definition
Correlation is used to test relationships between quantitative variables or categorical variables. In other words, it’s a
measure of how things are related. The study of how variables are correlated is called correlation analysis.
Some examples of data that have a low correlation (or none at all):
Correlations are useful because if you can find out what relationship variables have, you can make predictions about
future behavior. Knowing what the future holds is very important in the social sciences like government and healthcare.
Businesses also use these statistics for budgets and business plans.
Correlation coefficient formulas are used to find how strong a relationship is between data. The formulas return a value
between -1 and 1, where:
1 indicates a strong positive relationship.
-1 indicates a strong negative relationship.
A result of zero indicates no relationship at all.
Correlation coefficient formulas are used to find how strong a relationship is between data. The formulas return a
value between -1 and 1, where:
www.knowledgearena.in Page 36
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
Regression:
What is Regression?
If you’re just beginning to learn about regression analysis, a simple linear is the first type of regression you’ll
come across in a stats class.
Linear regression is the most widely used statistical technique; it is a way to model a relationship between two sets
of variables. The result is a linear regression equation that can be used to make predictions about data.
Linear” means line. The word Regression came from a 19th-Century Scientist, Sir Francis Galton, who coined the term
“regression toward mediocrity” (in modern language, that’s regression toward the mean). He used the term to describe the
phenomenon of how nature tends to dampen excess physical traits from generation to generation (like extreme height).
www.knowledgearena.in Page 37
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
independent variable X against one dependent variable Y. Technically, in regression analysis, the independent variable is
usually called the predictor variable and the dependent variable is called the criterion variable. However, many people just
call them the independent and dependent variables. More advanced regression techniques (like multiple regression) use
multiple independent variables.
Regression analysis can result in linear or nonlinear graphs. A linear regression is where the relationships between
your variables can be described with a straight line. Non-linear regressions produce curved lines.(**)
www.knowledgearena.in Page 38
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
By random selection we mean a selection where each and every item of the population has an equal chance of being
selected in the sample. In other words, the selection must not be made by deliberate exercise of one’s discretion. A sample
selected in this manner would be a representative of the population. If this condition is satisfied it is possible for one to
depict fairly accurately the characteristics of the population by studying only a part of it. Hence, this law is of great
practical significance because it makes possible a considerable reduction of the work necessary before any conclusion is
drawn regarding a large universe. For example, if one intends to make a study of the average height of the students of an
University it is not necessary to measure the heights of each and every student. A few students may be selected at random
from each college, their heights may be measured and the average height of university students in general may be inferred.
It should be noted that the results derived from sample data may be different from that of the population. This is for the
simple reason that the sample is only a part of the whole universe. For example, the average height of students of the
University may come out to be 160cm. By census method whereas it may be 159cm. or 161cm. for the sample taken. It
should be just a coincidence if the height comes out to be exactly 160cm. under both the methods. However, there would
not be much difference in the results derived if the sample is a representative of the population.
www.knowledgearena.in Page 39
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
8 heads and 2 tails, or 7 heads and 3 tails. If the same experiment is carried out 1,000 times the chance of 500 heads and
500 tails would be very high, i.e., the result would be very near to 50% heads and 50% tails. The basic reason for such
likelihood is that the experiment has been carried out sufficiently large number of times and possibility of variation in one
direction compensating others in a different direction is greater. If at one time we get continuously 5 heads, it is likely that
at other time we may get continuously 5 tails, and so on, and for the experiment as a whole the number of heads and tails
may be more or less equal. Similarly, if it is intended to study the variation in the production of rice over a number of
years and data are collected from one or two States only, the result would reflect large variations in production due to the
favourable factors in operation. If, on the other hand, figures of production are collected for all the States in India, it is
quite likely that we find little variation in the aggregate. This does not mean that the production would remain constant for
all the years. It only implies that the changes in the production of the individual States will be counterbalanced so as to
reflect smaller variations in production for the country as a whole.
www.knowledgearena.in Page 40
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
Principle of persistence of small numbers: According to this principle if some of the items in a population
possess markedly distinct characteristic from the remaining items then this tendency would be revealed in the sample
value also rather this tendency of persistence will be there even if the population size is increased or even in the case of
large sample.
Principle of validity: A sample design is termed as valid if it enables us to obtain valid tests & estimates about the
population parameters
Principle of optimization: this principle stresses the need of obtaining optimum results in terms of efficiency cost of
the sample design with the source available at our disposal.
www.knowledgearena.in Page 41
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
www.knowledgearena.in Page 42
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
5. If one regression coefficient is greater than unity, then others will be lesser than unity.
6. The geometric mean between the two regression coefficients is equal to the correlation coefficient
o R=sqrt(byx*bxy)
Also, the arithmetic means (am) of both regression coefficients is equal to or greater than the coefficient of
correlation.
(byx + bxy)/2= equal or greater than r.
7. The regression coefficients are independent of the change of the origin. But, they are not independent
of the change of the scale. It means there will be no effect on the regression coefficients if any constant is
subtracted from the value of x and y. If x and y are multiplied by any constant, then the regression
coefficient will change.
Range:-
The Range is the difference between the lowest and highest values.
Example: In {4, 6, 9, 3, 7} the lowest value is 3, and the highest is 9.
So the range is 9 − 3 = 6.
Probable Error of Correlation Coefficient:-
Definition: The Probable Error of Correlation Coefficient helps in determining the accuracy and reliability of the
value of the coefficient that in so far depends on the random sampling.
In other words, the probable error (P.E.) is the value which is added or subtracted from the coefficient of correlation (r) to
get the upper limit and the lower limit respectively, within which the value of the correlation expectedly lies.
The probable error of correlation coefficient can be obtained by applying the following formula:
www.knowledgearena.in Page 43
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
r = coefficient of correlation
N = number of observations
There is no correlation between the variables if the value of ‘r’ is less than P.E. This shows that the coefficient
of correlation is not at all significant.
The correlation is said to be certain when the value of ‘r’ is six times more than the probable error; this shows
that the value of ‘r’ is significant.
By adding and subtracting the value of P.E from the value of ‘r,’ we get the upper limit and the lower
limit, respectively within which the correlation of coefficient is expected to lie. Symbolically, it can be expressed
Important Formulas:-
www.knowledgearena.in Page 44
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
www.knowledgearena.in Page 45
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
www.knowledgearena.in Page 46
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
www.knowledgearena.in Page 47
KNOWLEDGE ARENA FOR UGC NTA NET/JRF
www.knowledgearena.in Page 48