Additional Textual Learning Material - B5
Additional Textual Learning Material - B5
Notes
Structure:
3.1 Introduction
3.2 Sample Correlation Analysis
3.3 Types of Correlation
3.4 Methods of Determining Correlation
3.5 Scatter Diagram
3.4 Methods of Determining Correlation
3.6 Karl Pearsons Co-efficient of Correlation
3.7 Rank Correlation
3.8 Regression Analysis
3.9 Time Series Analysis
3.10 Trend Analysis
3.11 Components of Time Series
3.12 Methods of Finding Trend
3.13 Moving Average Methods
3.14 Least Squares Methods
3.15 Summary
3.16 Check Your Progress
3.17 Questions and Exercises
3.18 Key Terms
3.19 Check Your Progress: Answers
3.20 Case Study
3.21 Further Readings
3.22 Bibliography
Objectives
Tippett states that, “the effect of correlation is to reduce the range of uncertainty Notes
of our prediction.”
In the words L. R. Conner “if two or more quantities vary in sympathy so that
movements in one tend to be accompanied by corresponding movement in the other then
they are said to be correlated.”
The statistical technique which deals with the association between two or more
variables is known as correlation analysis.
A. M. Tuttle defines correlation analysis as the Co-variation between two or more
variables.
Uses of Correlation
Advantages of Correlation
1. Correlation research allows researchers to collect much more data than
experiments.
2. Correlation research is that it opens up a great deal of further research to other
scholars.
3. It allows researchers to determine the strength and direction of a relationship
so that later studies can narrow the findings down and if possible, determine
causation experimentally.
4. Gain quantitative data which can be easily analysed.
5. No manipulation of behaviour is required.
6. The correlation coefficient can readily quantify observational data.
Disadvantages of Correlation
1. Correlation research only uncovers a relationship; it cannot provide a conclusive
reason for why there's a relationship.
2. A correlative finding doesn't reveal which variable influences the other. For
example, finding that wealth correlates highly with education doesn't explain
whether having wealth leads to more education or whether education leads to
more wealth.
3. Reasons for either can be assumed, but until more research is done, causation
can't be determined.
Notes 4. Here, a third, unknown variable might be causing both. For instance, living in
the City of Bangalore can lead to both wealth and education.
1. Positive Correlation
The correlation is said to be positive correlation if the values of two variables changing
with same direction.
Example: Production expenses and sales, Height and weight, water consumption
and temperature, study time and grades etc.
2. Negative Correlation
The correlation is said to be negative correlation when the values of variables change
with opposite direction. Example: Price and Quantity demanded, alcohol consumption and
driving ability etc.
Direction of the Correlation
Positive relationship – Variables change in the same direction.
As X is increasing - Y is increasing.
As X is decreasing - Y is decreasing.
Example, As height increases, so does weight.
Negative relationship – Variables change in opposite directions.
As X is increasing - Y is decreasing.
As X is decreasing - Y is increasing.
Example, As TV time increases, grades decrease.
3. Partial Correlation
In partial correlation more than two variables are recognised but only two variables
influence each other, the effect of other influencing variable is kept constant. In the above
example, if we limit our correlation analysis of yield and rainfall keeping fertilizer variable
as constant to becomes a problem of partial correlation.
1. Linear Correlation
Correlation is said to be linear when the amount of change in one variable tends
to bear a constant ratio to the amount of change in the other. The graph of the variables
having a linear relationship will form a straight line.
Example:
X = 1, 2, 3, 4, 5, 6, 7, 8
Y = 5, 7, 9, 11, 13, 15, 17, 19
Y = 3 + 2x
2. Non Linear correlation
The correlation would be non linear if the amount of change in one variable does
not bear a constant ratio to the amount of change in the other variable
X 1 2 3 4 5
Y 7 14 21 28 35
Notes 6. 40 70 y
Scatter
70 Diagram
7. 60 50
60
8. 20 40
50
Statistics
9. 30 60 Estimating Line
40
10. 50 30
30
11. 10 30 Low Degree of
20
Positive
12. 20 10 Correlation
10
0 x
10
20
30
40
50
60
70
Advantages of Scatter Diagram Maths
xy
When deviation taken from actual mean: r =
x 2 y2
( x )( y)
xy
n
When deviation taken from an assumed mean: r = ( x ) 2 ( y ) 2
x 2 y 2
n n
Co-efficient of Determination
r 2 = 0.81 this would mean that 81% of the variation in the dependent variable has
been explained by the independent variable.
The maximum value of r 2 is 1 because it is possible to explain all of the variation
in y but it is not possible to explain more than all of it.
Coefficient of Determination = Explained variation / Total variation.
Probable error of Co-efficient of Correlation gives us the two limits within which the
co-efficient of correlation of Series selected at random from the same universe is likely
to fall. The formula for the probable error of r is as follows.
1 r2
0.6745
N
Illustration: 1
1 r2 1 (0.6 )2
PE = 0.6745 = 0.6745 = 0.6745 0.08 = 0.06
N 64
Illustration: 2
For a given distribution the value of correlation is 0.64 and its probable error is
0.13274. Find the number of items in the series.
Solution: Notes
2
1r
P.E. = 0.6745 x Given, PE = 0.13274 Correlation (r) = 0.64
n
n = ?
1 ( 0.64 ) 2
PE = 0.6745 x
n
0.6745(1 0.4096)
0.13274 = n 0.13274 = 0.6745 x 0.5904
n
0.3982
n = 0.13274 n = 2.999 n = 9 (Approx)
Illustration: 3
Compute Karl Pearson’s Co-efficient of Correlation from the following data and calculate its
probable error and interpreter result.
Marks in Accountancy 77 54 27 52 14 35 90 25 56 60
Marks in English 35 58 60 40 50 40 35 56 34 42
Solution:
Calculation of Karl Pearson’s Co-efficient of Correlation
X X – 49 X2 Y y – 45 Y2 Xy
X Y
77 + 28 784 35 – 10 100 – 280
54 +5 25 58 + 13 169 + 65
27 – 22 484 60 + 15 225 – 330
52 +3 9 40 –5 25 – 15
14 – 35 1,225 50 +5 25 – 175
35 – 14 196 40 –5 25 + 70
90 + 41 1,681 35 – 10 100 – 410
25 – 24 576 56 + 11 121 – 264
56 +7 49 34 – 11 121 – 77
60 + 11 121 42 –3 9 – 33
xy 1449
r = 2 2 = 0.666
x y 5150 920
1 r2 1 (0.666)2 0.3753
PE = 0.6745 = 0.6745 = 0.119
N 10 3.162
Notes Illustration: 4
The following data relate to age of employees and the number of days they were reported
sick in a month.
Age 30 32 35 40 48 50 52 55 57 61
Sick days 1 0 2 5 2 4 6 5 7 8
Calculate Karl Pearson’s Co-efficient of Correlation
Solution:
Calculation of Karl Pearson’s Co-efficient of Correlation
Age x– x X2 Sick days y– y y2 xy
x x y y
30 – 16 256 1 –3 9 + 48
32 – 14 196 0 –4 16 + 56
35 – 11 121 2 –2 4 + 22
40 –6 36 5 +1 1 –6
48 +2 4 2 –2 4 –4
50 +4 16 4 0 0 0
52 +6 36 6 +2 4 + 12
55 +9 81 5 +1 1 +9
57 + 11 121 7 +3 9 + 33
61 + 15 225 8 +4 16 + 60
= 460 x2 = 1092 y = 40 y2 = 64 xy = 230
460
x = 46
10
40
y = 4
10
xy 230 230
r = 0.87
2 2 1092 64 263.36
x y
There is a high degree of positive Correlation between age and number of days reported sick.
Illustration: 5
Solution: Notes
Calculation of Co-efficient of Correlation
X X – 34 x2 Y y – 35 y2 XY
x
48 + 14 196 45 + 10 100 + 140
35 +1 1 20 – 15 225 – 15
17 – 17 289 40 +5 25 – 85
23 – 11 121 25 – 10 100 + 110
47 + 13 169 45 + 10 100 + 130
x = 170 x2 = 776 y = 175 y2 = 550 xy = 280
x 170 y 175
x = 34 y = 35
N 5 N 5
xy 280 280
r = 2 2 = = 0.429
x y 776 550 653.299
Illustration: 6
From the following data of six commodities calculate Karl Pearson’s Co-efficient of correlation
between sales and price per kg sold.
Commodities Sales (Kgs) Total Price
Rice 100 8,000
Wheat 120 10,000
Sugar 150 13,000
Gram 100 10,000
Oil 160 19,200
Ghee 170 30,940
Solution:
In order to simplify the calculations the sales column is divided by 10 and price column is
divided by 1000.
Calculate of Coefficient of Correlation
X x = x – 15 x2 Y y = y – 15 y2 xy
10 –5 25 8 –7 49 39
12 –3 9 10 –5 25 15
15 0 0 13 –2 4 0
10 –5 25 10 –5 25 25
16 +1 1 19.2 4.2 17.64 4.2
17 +2 30.94 15.94 15.94 254.08 31.88
x = – 10 x2 = 64 y = 1.14 y2 = 374.72 xy = 111.8
Notes
( x )( y)
xy ( 10)(114
. )
n 111.08
6
r = ( x ) 2 ( y ) 2 =
x 2 y 2 ( 10)2 . )2
(114
n n 64 374.72
6 6
112.98 112.98
= = 10.84
4733 374.50 133.14
As there is a high positive correlation between sales and total price it means when
Sales is high, the total Price will also be high.
Illustration: 7
From the data given below, compute Karl Pearson’s co-efficient of correlation.
X series Y series
Number of items 15 15
Arithmetic mean 25 18
Square of deviation from Arithmetic mean 136 138
Summation of products of deviations from the arithmetic means of X and Y-122.
Solution:
x = 25, x2 = 136 xy = 122
y = 18 y2 = 138 n = 15
( x )( y) 25 x 18
xy 122
n 15
r = ( x )2 ( y) 2 = (25)2 (18)2
x 2 y 2 136 138
n n 15 15
122 30 92 92
= = (9.71)(10.78) 104.67 = 0.8789
136 41.66 138 21.6
Karl Pearson’s co-efficient of correlation is 0.8789
Illustration: 8
Calculate Karl Pearsons co-efficient of correlation for the following data regarding
price and demand of a certain commodity.
Price (in `) 21 22 23 24 25 26 27 28 29
Demand (in ‘000 units) 20 19 19 17 17 16 16 15 14
Solution:
A = 25 A = 17
21 –4 16 20 3 9 –12
22 –3 9 19 2 4 –6
23 –2 4 19 2 4 –4 Notes
24 –1 1 17 0 0 0
25 0 0 17 0 0 0
26 1 1 16 –1 1 –1
27 2 4 16 –1 1 –2
28 3 9 15 –2 4 –6
29 4 16 14 –3 9 –12
x = 0 60 y = 0 32
( x )( y )
xy
n 43 0 43
r = 2 ( x ) 2 2( y )2 = = 60 x 32
x y 60 0 32 0
N N
43 43
= = = –0.981
1920 43 .81
Illustration: 9
% of pass % of fail x = x – a x2 y = y – b y2 x y
(x) (y) a = 50 b = 50
405 195
X = 6
= 67.5 =
6
= 32.5
N .xy x . y
r =
Nx x ) 2
2
Ny 2 ( y ) 2
Notes
19,350 11,025 8,325 8,325
= 19,350 11,025 19,350 11,025 = 2 = = –1
( 8,325) 8,325
PE = 0.6745
1 r2
=
0.6745 1 ( 1)2 =
0.6745 x 0
= 0
n 6 6
6 D2
rn = 1
N ( N 2 1)
Where, r = Rank correlation
D = Difference of rank between paired items in two series
N = Total number of observation
When equal ranks the assigned to some entries an adjustment in the usual formula
for calculating the rank coefficient of correlation is made. The adjustment consists of adding
1/12 (m2 – m) to the value of d 2 , where ‘m’ Stands for the number of items whose
ranks are common. If there are more than one such group of items with common rank,
this value is added as many times, the number of such groups. The formula can this be
written
RS
6 d2
1
(m3 m)
1
(m3 m)
UV
rn = 1 T 12 12 W
N3 N
Amity Directorate of Distance and Online Education
Forecasting Techniques 109
Illustration: 1
From the following data calculate co-efficient of rank correlation between X and Y.
X: 36 56 20 65 42 33 44 50 15 60
Y: 50 35 70 25 58 75 60 45 80 38
Solution:
Calculation of co-efficient of Rank Correlation
X R1 Y R d= R -R d2
2 1 2
36 7 50 6 1 1
56 3 35 9 -6 36
20 9 70 3 6 36
Amity Directorate of Distance and Online Education
110 Quantitative Techniques in Management
Notes 65 1 25 10 -9 81
42 6 58 5 1 1
33 8 75 2 6 36
44 5 60 4 1 1
50 4 45 7 -3 9
15 10 80 1 9 81
60 2 38 8 -6 36
d=0 d2=318
Illustration: 2
80 123 4 4 0 0
91 135 2 2 0 0
99 154 1 1 0 0
71 110 5 6 –1 1
61 105 7 8 –1 1
81 134 3 3 0 0
70 121 6 5 1 1
59 106 8 7 1 1
n = 4
6d2 6(4) 24
r = 1 – = 1 – = 1 –
2
n(n 1) 8(82 1) 8(63)
24
= 1 – = 1 – 0.0476
504
r = 0.9524
Illustration: 3
Solution: Notes
Computation of Rank Correlation co-efficient
X Y R1 R2 d = R1 – R2 d2
60 75 1 1 0 0
34 32 8 7 1 1
40 34 7 5 2 4
50 40 2 3 –1 1
45 45 3 2 1 1
41 33 6 6 0 0
22 12 9 9 0 0
43 30 4 8 –4 16
42 36 5 4 1 1
d2 = 24
1 6d2 6 x 24 144
r = = 1 1 = 1 – 02 = 0.8
3
n n 93 9 720
Illustration: 4
Calculate the co-efficient of rank correlation for the following data of marks obtained by 9
students in Accountancy and Statistics at an Examination.
Notes 15 2 9 9 0 0
60 5 4 8 - 4 16
d2=45.5
3 3
(m1 m) (m2 m)
6 d2
12 12
Rank correlation co-efficient = r 1
s
n3 n
Where
(33 3) (23 2) m no. of times each
6 45.5
12 12
1 Valuerepeats
3
9 9
n totalno. of values
Illustration: 5
Calculate the rank correlation between the length of service and order of merit from the
following data:
Employee: A B C D E F G H I J K L
Yrs of service: 5 2 10 8 6 4 12 2 7 5 9 8
Order of Merit: 6 12 1 9 8 5 2 10 3 7 4 11
(Efficiency)
Solution:
Years of service should be ranked, whereas the order of merit not.
d = 0 d2 = 111.5
Tied Ranks - (I) In X series (a) 8 th and 9th and (b) 11th and 12th and (c) 4th and 5th
1 1 1 Notes
6[d2 (23 2) (23 2) (23 2)]
rs 1 12 12 12
n3 n
Regression Co-efficient
The rate of change of Variable for unit change in the other variable is called the
regression co-efficient of former on the latter. Since there are two regression lines there
are two regression co-efficients. The rate of change of x for unit change in y is called
regression coefficient of x on y. It is the Co-efficient of y in the regression equation. When
it is in the form of x = n + by. This is denoted by bxy
x
bxy = r
y
x
Similarly, regression Co-efficient of y on x is denoted by byx = r
y
Notes (2) It predicts the values of dependent variables from the known values of
independent variables.
(3) It is used in Statistical estimation of demand curve, supply curve, Production
function, Cost function, consumption function etc.
(4) To study correlation with the help of regression.
Regression Lines
If we take two variables X and Y we shall have two regression lines as the regression
of X on Y and the regression of Y on X. The regression line of Y on X gives the most
probable values of Y for given value of X and the regression line of X on Y gives the most
probable values of X for given values of Y. However, When there is either perfect positive
or negative correlation between the two variables, the regression lines will coincide, i.e.,
we will have only one line.
The Regression Line is the line that best fits the data, such that the overall distance
from the line to the points (variable values) plotted on a graph is the smallest. In other
words, a line used to minimize the squared deviations of predictions is called as the
regression line.
The regression line gives the average relationship between the two variables in
mathematical form. The Regression would have the following properties:
a) (Y – Yc) = 0 and
b) (Y – Yc)2 = Minimum
The farther the two regression lines from each other, the lesser is the degree of
correlation and the nearer the two regression lines to each other, the higher is the degree
of correlation. If the variables are independent, r is zero and the lines of regression right
angles, i.e., parallel to ox and oy.
For two variables X and Y, there are always two lines of regression –
Regression line of X on Y:
gives the best estimate for the value of X for any specific given values of Y?
X = a + bY
a = X - intercept
b = Slope of the line
X = Dependent variable
Y = Independent variable
For two variables X and Y, there are always two lines of regression –
Regression line of Y on X:
gives the best estimate for the value of Y for any specific given values of X
Y = a + bx
a = Y - intercept
b = Slope of the line
r byx bxy
If byx is positive than bxy should also be positive and vice versa.
If one regression coefficient is greater than one the other must be less than one.
The coefficient of correlation will have the same sign as that our regression coefficient.
Arithmetic mean of byx and bxy is equal to or greater than co-efficient of correlation.
byx + bxy / 2 r Regression coefficient are independent of origin but not of scale.
Regression equations are algebraic expressions of the regression lines. Since there
are two regression lines, There are two regression equations – The regression equation
of X on y is used to describe the variations in the values of X for given changes in y and
the regression equation of y on x is used to describe the variation in the values of y for
given changes in x
The two Regression Equations
The two regression equations are as follows:
(1) Regression equation of x on y
Y = a + bx
The Values of a and b obtained by solving the two simultaneous equation
y = Na + bx
xy = ax + bx2
(2) Regression equation of x on y
X = a + by
The values of a and b are obtained by solving the two simultaneous equations
x = Na + by
y = ay + by2
Illustration: 1
Illustration: 2 Notes
State the relationship between Correlation co-efficient of two variables and their
regression co-efficient of two variables and their regression co-efficients.
Regression Co-efficient of x on y =
x
bxy = r y
Regression Co-efficient of y on x
y
byx = r
x
1 y
bxy byx = r r
y x = r2 r
r= bxy. byx
It can be Said that correlation Co-efficient is the geometric mean of regression
Co-efficients.
Illustration: 3
(x - 20) = 0.75
2
3
by 25 g = 0.50 y
2
3
by 25 g
= 0.50y + 20 – 12.5 = 0.50y + 7.5
The regression equation of y on x is
y 3
( y y ) = r ( x x ) (y – 25) = 0.75 ( x 20)
x 2
= 1.125n + 25 – 22.5 y = 1.125n + 2.5
The equations are
x = 0.5y + 7.5
x = 1.125x + 2.5
Notes Illustration: 4
You are given the following information about advertisement and Sales of a Company
Mean 10 90
Variance 9 16
Correlation Co-efficient = 08
(1) Calculate the two regression lines
(2) Find the likely Sales When advertising expenses are ` 15 lakhs and
(3) What should be the advertisement expenses
If the Company Wants to attain a Sales target of ` 190 Lakhs.
Solution:
Regression equation of x on y
x 3
( x x ) = r ( y y ) x – 10 = 0.8 ( y 90 )
y 4
x – 10 = 0.6y – 54 x = 0.6y – 44
The advertisement expenses for Sales Target of ` 190 lakhs will be
X = 0.6 190 – 44 = ` 70 Lakhs
Regression equation of y on x:
y 4
( y y ) = r ( x x ) y – 90 = 0.8 ( x 10 )
x 3
x – 90 = 1.0667x – 10667 y = 1.0667x + 79.333
Sales When advt expenses is ` 15 Lakhs
Y = 1.0667 15 = 79.333 = 95.33 Lakhs
Illustration: 5
Obtain the two regression equations by taking deviations from actual mean of x and
y Series
X 2 4 6 8 10
Y 5 7 9 8 11
Solution:
Calculation of Regression equations
X X–6=x X2 Y y–8=y Y2 Xy
2 –4 16 5 –3 9 12
4 –2 4 7 –1 1 2
6 0 0 9 1 1 0
8 2 4 8 0 0 0
10 4 16 11 3 9 12
x = 30 x2 = 40 y = 40 =0 y2 = 20 xy = 26
Amity Directorate of Distance and Online Education
Forecasting Techniques 119
x 30 y 40 Notes
x = N 5 6 y =
N
5
8
Regression equation of x on y:
xy 26
(x x ) = b( y y) b= = 1.3
y2 20
(x – 6) = 1.3 (y – 8) x – 6 = 1.3y – 10.4 x = 1.3y – 4.4
Regression equation of y on x:
xy
( y y) = b(x x ) b=
x2
26
= 0.65 (y – 8) = 0.65 (x – 6) y – 8 = 0.65x – 3.9
40
y = 0.65x + 4.1
Illustration: 6
y 20
( y y ) = r ( x x ) (y - 192) = 0.6 ( x 170 )
x 60
y – 192 = 0.2x – 34 y = 0.2x + 158
x 8500
x = N 50 Rs.170
y 9600
y = Rs.192
N 50
Given x = 200, y = ?
Y = 0.2 200 + 158 = 198
Illustration: 7
The following table shows the exports of New Cotton and the imports of manufactured
goods into India for 7 years
Exports (in Crores of `) 42 44 58 55 89 98 60
Imports (in crores of `) 56 49 53 58 67 76 58
Notes Obtain the two regression equations and estimate the imports when exports in a
particular year were to the value of ` 70 crores.
Solution:
‘X’ Series represents Export ‘Y’ Series represents Import
Calculation of Regression equations
58 –2 4 53 –7 49 14
55 –5 25 58 –2 4 10
89 29 841 67 7 49 203
60 0 0 58 –2 4 0
x 446
x = N 7 63.714
y 417
y = 59.571
N 7
Regression equation of x on y:
x x = bxy (y – y )
N dxdy ( dx )( dy)
bxy =
N dy 2 ( dy)2
7 1083 26 3 7581 78 7659
= 2 = = 2.1983
7 499 ( 3) 3484 3484
y y = byx ( x x )
7 1083 26 3 7659
= 2 = 0.3911
7 2894 (26) 19582
y – 59.571 = 0.3911(x – 63.714)
y – 59.571 = 0.3911x – 24.9185
y = 0.3911x + 34.6525
Calculate of y When x = 70
y = 0.3911 70 + 34.6525 = 62.0295 crores
Illustration: 8
The following data gives the demand (x) and Supply (y) in a market for 11 days
X 61 72 73 73 63 84 80 66 76 74 72
Y 40 52 59 53 61 58 56 42 58 50 50
Solution:
Calculation of Regression equation
X dx = x–72 dx2 Y dx = y – 52 dy2 dxdy
61 – 11 121 40 – 12 144 132
72 0 0 52 0 0 0
73 1 1 59 7 49 7
73 1 1 53 1 1 1
63 –9 81 61 9 81 – 81
84 12 144 58 6 36 72
80 8 64 56 4 16 32
66 –6 36 42 – 10 100 60
76 4 16 58 6 36 24
74 2 4 50 –2 4 –4
72 0 0 50 –2 4 0
x = 794 dx2 = 468 y = 579 dy = 7 dy2=471 dxdy=243
Regression equation of x on y:
N dxdy ( dx )( dy)
bxy =
N dy 2 ( dy)2
11 243 2 7 2659
= = 0.5181
11 471 (7)2 5132
x – 72.18 = 0.5181 (y – 52.64)
x – 72.18 = 0.5181 y – 27.2738
x = 0.5181y + 44.9062
y y = bys ( x x )
x 794
x = N 11 72.18
y 579
y = 52.64
N 11
Illustration: 9
The following data relate to the age of husbands and wives. Obtain the two regression
equations and determine the most likely age of husband when age of wife is 25 years
Age of Husbands 25 28 30 32 35 36 38 39 42 55
Age of Wives 20 26 29 30 25 18 26 35 35 46
Solution:
Age of Husband is represented by X Series and age of Wives is represented by Y Series
X X = x – 36 X2 Y Y = y – 29 Y2 Xy
25 – 11 121 20 –9 81 99
28 –8 64 26 –3 9 24
30 –6 36 29 0 0 0
32 –4 16 30 1 1 –4
35 –1 1 25 –4 16 4
36 0 0 18 – 11 121 0
38 2 4 26 –3 9 –6
39 3 9 35 6 36 18
42 6 36 35 6 36 36
55 19 361 46 17 289 323
x = 360 x2 = 648 y = 290 y2 = 598 xy = 494
x 360
x = N 10 36 years
y 290
y = 29 years
N 10
Time series analysis are basic to understanding past behaviour, evaluating current
accomplishments, planning for future operations and comparing different time series. Thus
a series of successive observations of the same phenomenon over a period of time are
called “time series”.
Definitions
Notes Y (temperature, closing price of share, etc.) at the time t1, t2 .... Thus, Y is a function
of t, symbolised by y = F(t).” – Spiegel
“A time series may be defined as a collection of magnetite belonging to different
time periods, of some variable or different time periods, of some variable or composite
of variables, such as production of steel, per capita income, gross national product, price
of tobacco or index of industrial production.” – Ya–Lun–Chou
The uses of time series is of well significance to the economist, scientist, sociologist,
biologist, researcher and businessman etc. The following uses are:
(i) It helps to understand past behaviour.
(ii) It helps in evaluating current accomplishment.
(iii) It helps in planing future operations.
(iv) It also helps in comparing the actual performance.
(v) It helps to study the factor which influence the changes in economic activities
and predict the future variations in them with certain limitations.
Fig. (i) Shows a graph of a long term or secular trend line. Fig. (ii) shows this long
term trend line with a superimposed cyclical movement. fig. (iii) shows a superposition
of a seasonal movement on the graph of figure (ii).
Amity Directorate of Distance and Online Education
Forecasting Techniques 125
The ideas presented above provide us with a possible technique for analysis time Notes
series. So, assume that the time series variable Y is a assume that the time series variable
Y is a product of the variables, T C S and I which produce respectively the trend, cyclical,
seasonal and irregular movements.
70
65
60
55
Num ber of professional
50
45
(in lakhs)
40
35
30
25
20
15
10
1996 1997 1998 1999 2000 2001 2002 2003 2004 2005
Year
Fig: IT professional of India source: Department of IT
Components
of
time serices
Notes Characteristic movement of time series may be classified into four main types, often
called components of a time series. The components of a time series are the various
elements which can be separated from the observed data. The components of time series
are classified as under:
The secular trend is the basic long–term tendency, of a particular activity to grow
or to decline. It indicates the presence of a factors that persist for a considerable duration
factors such as population changes, technological improvements, price level fluctuations
or various conditions that are peculiar to individual industries or establishments. Thus,
the tendency of data to grow or decline over a long period is technically called secular
movement or simple trend. For example in India, the sales of car is showing an increasing
trend and sales of black and white TV is showing declining trend.
“The secular trend is that characteristic of a time series which extends consistently
throughout the entire period of time under consideration”. – Henry L. Alder
2. Cyclical Analysis
Its refer to the long term swing about a trend line or curve. These cycles, as they
are sometimes called may or may not be periodic, i.e., they may or may not follow exactly
similar pattens after equal intervals of time. In business and economic activities,
movements are considered cyclical only if they recur after time intervals of more than a
year. An important example of cyclical movements are the so called business cycles
representing intervals of prosperity, recession, desprsion and recovery.
Up and down movements which are different from seasonal fluctuations, in that they
extend over longer period of time–usually two or more years.” – Linclon L. Choo.
3. Seasonal Analysis
Its refer to the sporadic motions of time series due to chance events such as floods,
strikes, elections, fires, earthquakes, unusual weather etc. Random variations in a time
series are variations which cannot be accounted for by the secular trend, seasonal
movements or cyclical fluctuations. So, it is ordinarily assumed that such events produces
variations lasting only short time, it is conceivable that they may be so intense as to Notes
result in new cyclical or other movements.
“Random or erratic variations are variations which occur in a completely unpredictable
fashion”. – Henry L. Alder.
200
180
160
Production (in lakhs)
140
Seasonal
120
100 Cyclical
Trend
80
60
40
20
1996 1997 1998 1999 2000 2001 2002 2003 2004 2005
Year
Example - 1
Using the method of Free hand determine the trend of the following data:
Year: 1998 1999 2000 2001 2002 2003 2004 2005
Production: 42 44 48 42 46 50 48 52
(in tonnes)
Solution:
Y
56
X axis 1cm = year
54 Y axis 2 cm = 2 tonce
52
Production
50
48
46 Trend line
44 Actual data
42
40
O X
1998 1999 2000 2001 2002 2003 2004 2005
Years
Which consists of fitting a trend line or curve simply by looking at the graph, can
be used to estimated trend.
The time is shown on the horizontal axis and the value of the variable on the vertical
axis. The fitting of the trend may be straight line or a curved line and it may be done
free hand by scale rules, spline, string or French curves of different shapes. Smooth out
irregularities by drawing a free hand curve through the scatter points.
Merits:
(i) A free hand trend fitting enables an understanding of the character of time series.
(ii) These are its flexibility and simplicity.
(iii) This method only used to describe all types of trends – linear and non linear.
Demerits:
(i) This is depending too much on individual judgment.
(ii) It does not involved any complex mathematical techniques.
(iii) It does not enable us to measure trend in precise quantitative terms.
(iv) Different trend curves could be obtained by different persons for the same data.
It is highly subjective.
Semi–Average Method
The method is used only when the trend is linear or almost linear. For non-linear
trends this method is not applicable. It is used for the calculation of averages, and averages
are affected by extreme values. Thus if there is some very large value or very small value
in the time series, that extreme value should either be omitted or this method should not
be applied.
Merits:
(i) It is simple and easy to understand.
(ii) It can compared with the moving average of the least squares method of
measuring trend.
(iii) This method is objectivity in the sense that it does not depend of the personal
judgment.
(iv) It is also applicable where the trend is linear or approximately linear.
Demerits:
(i) This method is based on the assumption that there is a linear trend which may
not be true.
(ii) It is not suitable when time period represented by average is small.
(iii) The use of arithmetic mean for obtaining semi–average may be questioned
because of limitation of the method.
Illustration - 1
Using the method of semi–average determine the trend of the following data:
Year : 1998 1999 2000 2001 2002 2003 2004 2005 Notes
Production: 48 42 60 54 38 66 70 66
Solution:
The number of observations are even i.e., 8. The two middle parts will be 1997 to
2001 and 2002 to 2006.
72
68
64
Production
60 –
X2 = 60
56
Trend line
52
–
X1 = 51 Actual line
48
44
40
O X
1998 1999 2000 2001 2002 2003 2004 2005
Years
Illustration - 2
Calculate trend values from the following data by the method of semi–average:
Year: 1996 1997 1998 1999 2000 2001 2002 2003 2004 2005
Sales: 30 35 37 24 42 36 27 45 40 42
Notes Solution:
Year Sales Averages Annual changes Trend value
1996 30 29.2
1997 35 30
170
1998 37 34 30.8
5
1999 24 31.6
2000 42 0.8 32.4
2001 36 33.2
2002 27 .... 34
2003 45 34.8
2004 40 35.6
2005 42 36.4
Let us have two periods of 5 years each 1996 to 2000 and 2001 to 2005. The averages
for the two periods are:
30 35 39 24 42 170
34.......... x1
5 5
36 27 45 40 42 190
and 38.......... x2
5 5
The increase of two averages is 38 – 34 = 4 which takes place in 5 years, therefore,
annual change is .... = 0.8
“Graph showing Sales with Trend line”
Y
47
45
43
41
–
39 X2
37 Trend line
Sales
35
33 –
X1 Actual line
31
29
27
25
O X
1996 1997 1998 1999 2000 2001 2002 2003 2004 2005
Years
Illustration - 3 Notes
Draw a trend by the method of semi–average from the following data:
Year: 1998 1999 2000 2001 2002 2003 2004 2005
Sales (‘000 units): 195 100 104 90 95 102 110 116
Also predict the Sales fort he year 2003 from the graph.
Solution:
Year Sales(x) Semi Average
1998 195
489
1999 100 = 122.25 ....... x1
4
2000 104
2001 90
2002 95
423
2003 102 .... = 1.5.75 ..... x2
4
2004 110
2005 116
140
2003 Sales = 108,00 units
130
120 –
X1
Trend line
110
– Actual line
100 X2
90
X
O
1998 1999 2000 2001 2002 2003 2004 2005
Years
Notes covering a fixed time interval, called moving average period. It is shown against the centre
of the period.
The moving average for period ‘t’ is a series of successive averages of values at a
time, starting with 1st, 2nd and 3rd to ‘t’ terms. Here, the first average is the mean of
the 1st to ‘t’ terms, the second is the mean of the ‘t’ terms from 2nd to (t + 1) th terms
and third is the men of the 3rd to (t + 2)th terms and so on.
Hence, the time series values X1, X2, X3 .....
for different time periods, the moving average of period ‘t’ is given by:
1
1st value moving average = (X1 + X2 + .... Xt),
t
1
2nd value moving average = (X2 + X3 + ...... Xt + 1) and 3rd
t
1
value moving average = (X + X4 + ..... Xt + 2)
t 3
(a) Odd period: When the period is odd, if the period ‘t’ of the moving average
is odd the successive value of the moving averages are placed against the middle
value of concerned group of items. If t = 5, the first moving average value is
placed against the middle period i.e. third value and the second moving average
value is placed against the time period four and so on.
(b) Even period: When the period is even, it the period ‘t’ of moving average is
even there are two middle periods and the moving average value is placed
between the two middle terms of the time intervals. In the particular period t
= 4, the first moving average is placed against the middle of second and third
values, the second moving average is placed in between third and fourth values
and so on.
For odd: In the below table (a) have considered 5 years moving averages.
1995 50.0
1996 36.5
1997 43.0 212.9 42.6
1998 44.5 201.0 40.2
1999 38.9 197.1 39.4
2000 38.1 192.8 39.6
2001 32.6 190.0 38.0
2002 38.7 192.2 38.4
2003 41.7 187.9 37.6
2004 41.1
2005 33.8
Amity Directorate of Distance and Online Education
Forecasting Techniques 133
Here, the first moving total 219.9 of column 3 is the sum of the 1st through 5th entries Notes
of column 2. The second moving total 201.0 is the sum of the 2nd through 6th entries
in column 2 etc.
In practice, after obtaining the first moving total 212.9, the second moving total is
easily obtained by subtracting 50.0 (1st entry of column 2), the result being 201.0.
Succeeding moving totals are obtained similarly. Dividing each moving total by 5 yields
the required moving average.
In the below table ... have considered 4 years moving averages.
For even
Year Data 4 years 4 years
moving total moving total
1995 50.0
1996 36.5
1997 43.0 174.0 43.5
1998 44.5 162.9 40.7
1999 38.9 164.5 41.1
2000 38.1 154.1 38.5
2001 32.6 148.3 37.1
2002 38.7 151.1 37.8
2003 41.7 154.1 38.5
2004 41.1 155.3 38.8
2005 33.8
Here, the 4 year moving totals are obtained as in past (a), except that 4 entries
of column 2 are added instead of 5. Note that the moving totals are centred between
successive years, unlike part (a). This is always the case when an even number of years.
is taken in the moving average. If we consider that 1996, for example, stands for July
1, 1996, the first 4 year moving total is centred at Jan 1, 1997 or Dec. 31, 1996. The
4 year moving averages are obtained by dividing the 4 year moving totals by 4.
Merits:
(i) It is very simple to understand and easy to calculate as compared to other
methods.
(ii) It can be used in all facts of time series analysis, for the measurement of trends
as well as in connection with seasonal cyclical irregular components.
(iii) It calculate is simple because no higher degree mathematical calculations.
(iv) It considers all the values in the series. The extreme values are included in the
process of determining averages.
(v) It is a objective method. No personal judgment like freehand method.
Demerits:
(i) In this method, that data at the beginning and end of a series are lost.
(ii) It may generate cycles or other movements which were not present in the original
data.
Notes (iii) Its are strongly affected by extreme values of items. They are said to be sensitive
to.... movement in the data.
(iv) It does not establish functional relationship between the period and the value of
variables.
(v) It cannot determine irregular variations completely.
Illustration - 1
I II III IV V
Year Production Column of 3 Year 3 Year
differences moving total moving average
1999 112
396
2000 138 154 – 112 = 42 .....396 = 132
3
438
2001 146 170 – 138 = 32 ..... 438 = 146
3
470
2002 154 183 – 146 = 37 .... 470 = 157
3
507
2003 170 190 – 154 = 36 ..... 507 = 169
3
543
2004 183 .... 543 = 181
3
2005 190
Calculating total production of first three years which is (112 + 138 + 146 = 396). This
total is placed in column (IV) of three yearly moving total before the middle year. 2000 and in
column (V) the three year moving average 396/3 = 132 is placed before the year 2000 as shown
in the above table. And find the second moving total we find the difference of the production
of a year 2002 and 1999 which is (154 – 112 = 42). This is written in column (III) in front of
the year 2000 and the second moving total is (396 + 42 = 438). Such way it can complete
columns (III) and (IV) of the above table.
Illustration - 2
Find trend values from the following data using three yearly moving averages and show
the trend line on the graph.
Year Price (`) Year Price (`)
1994 52 2000 75
80
Scale :
X axis : 1cm = 1 year
77
Y axis : 1cm = 3 Rs.
74
71
68
Price
65
62
Trend line
59 Actual line
56
53
50
X
O
94 95 96 97 98 99 00 01 02 03 04 05
Years
Notes Illustration - 3
1995 21 - -
1996 22 66 22
1997 23 69 23
1998 24 72 24
1999 25 75 25
2000 26 78 26
2001 27 79 26.33
2002 26 - -
Y
Graphical Representation
28
X axis 1 year = 1 cm
27 Y area 1 unit = 1 cm
26
25
24
23
22
21
20
X
95 96 97 98 99 00 01 02
Illustration - 4
Find the four yearly moving averages for the following data:
Year: 1991 1992 1993 1994 1995 1996 1997 1998
Values: 30.1 45.4 39.3 41.4 42.2 46.4 46.6 49.2
Amity Directorate of Distance and Online Education
Forecasting Techniques 137
Solution: Notes
Year Values 4 yearly 4 yearly Centeral value
moving Total moving Average
1991 30.1
1992 45.4
156.2 39.05
1993 39.3 40.56
168.3 42.07
1994 41.4 42.195
169.3 42.32
1995 42.2 43.235
176.6 44.15
1996 46.4 45.125
184.4 46.1
1997 46.6
1998 49.2
Illustration - 5
Using four yearly moving averages determine the trend values and also plot the original and
trend values on a graph.
1994 75 2000 96
1985 62 2001 128
1996 76 2002 116
1997 78 2003 76
1998 94 2004 102
1999 84 2005 168
Solution:
1994 75
1995 62 291 72.75 75.125 1996
1996 76 310 77.50 80.25 1997
1997 78 332 83.00 85.50 1998
1998 94 352 88.00 94.25 1999
1999 84 402 100.50 103.25 2000
2000 96 424 106.00 105.00 2001
2001 128 461 104.00 104.75 2002
2002 116 422 105.50 110.50 2003
Graph showing production and trend value for the years 1994-2005
Scale: OX= 1 cm = 1year, OY = 1 cm = 1000 units.
Actual Line
Trend Line
Years
Illustration - 6
Calculate the trend values by five yearly moving average method and plot the same on a
graph from the following:
Year: 1995 1996 1997 1998 1999 2000 2001 2002 2003 2004 2005
Sales: 36 42 54 72 66 60 48 75 78 102 93
(’000’units)
1995 36 –
1996 42 –
1997 54 270 54
1998 72 294 58.8
1999 66 300 60
2000 60 321 64.2
2001 48 327 65.4
2002 75 363 72.5
Amity Directorate of Distance and Online Education
Forecasting Techniques 139
110
100 Scale :
X axis : 1cm = 1 year
Y axis : 1cm = 10,000 units
90
Sales (1000 units)
80
70
60
50
Actual line
40 5 yearly moving
average line
30
X
O
95 96 97 98 99 00 01 02 03 04 05
Years
Notes
Y
|Y–Ye|
|Y–Ye| Y
Ye
Ye
Y
X
O X1 X2 X3 X4
The trend line, the which is a “best fit” to a scatter diagram of n points, its employ
a theorem from elementary mathematics.
From the below figure, the equation of any no–vertical line can be written in the form
Y = a + bx,
where, a is the Y–intercept
and b is the slop of the line.
The slope is given b = tan , where is the angle measured from the positive X
– axis to the line and is positive or negative as is obtuse.
X
O
1999 –3 1998 –7
2000 –2 1999 –5
y xy
a = ; b=
n x2
Finally the trend line, the line of ‘best fit’ will be drawn under the least squares method.
It is the line from which the sum of the squares of the items, measured parallel to the
Y axis, is the least.
So equation represented by
Y = a + bx
least square method for Trend values
Yc = a + bx
y xy
a = ; b =
n x2
Here, x = independent variable (Deviation)
y = Dependent variable on x.
a = The ‘y’ intercept.
b = Indicate slope and signifies the changes in ‘x’ ...constant.
n = Number of observations.
xy = Product of x and y.
x2 = Square of x.
y = Sum of ‘y’ values
xy = Sum of th<e product of x and y values.
x2 = Sum of square value of x.
Merits:
(i) It is objective method which does not variations in the results.
(ii) It is very easy calculation because no higher degree calculation.
(iii) It determines the trend values and also reflects light on the seasonal, cyclical
and irregular variations.
Amity Directorate of Distance and Online Education
142 Quantitative Techniques in Management
Notes (iv) Since it is based on an algebraic method of calculating the trend, it is free from any
bias, subjectivity does not enter into it.
(v) It is a flexible method, the trend values calculated any period and answer should be
accurate.
Demerits:
(i) This method is in appropriate for a very short series and is unnecessary for along one.
(ii) It does not establish functional relationship between the period (x) and the values of
variable ‘y’.
(iii) It is a tedious method and involves more calculations compared to the methods
discussed earlier.
(iv) It can estimate a value only immediate future and not for distant future.
Illustration - 1
Fit a straight line trend by the method of least squares, tabulate the trend values and show
the values on a graph form the following figure of production of sugar factory:
Year: 2007 2008 2009 2010 2011 2012 2013
Production: 80 90 92 83 94 99 92
(‘000’ tons)
Solution:
y 630 xy 56
Here, a = 90 and b 2
n 7 x 28
Y
110
Scale :
105 X axis : 1cm = 1 year
Y axis : 1cm = 5000 tonnes
100
Production
95
90
Trend line
85
Actual line
80
X
O
99 00 01 02 03 04 05
Years
Illustration - 2
Fit a straight line trend to the following data taking X as the independent variable and prove
S(Y – Yc) = 0.
X: 2000 2001 2002 2003 2004 2005
Y: 1 1.8 3.3 4.5 6.3 10
Solution:
Calculation of Trend by the method of least squares
Year y x x2 xy Trend value
Yc = a + bx
2000 1 –5 25 –5 0.223
2001 1.8 –3 9 –5.4 1.927
2002 3.3 –1 1 –3.3 3.631
2003 4.5 +1 1 4.5 5.335
2004 6.3 +3 9 18.9 7.037
2005 10 +5 25 50 8.743
y 26.9
a= 4.483
n 6
xy 59.7
b= 0.852
x 2 70
y xy
a= and b =
n x 2
619 141
a= = 88.43 b= = 5.04
7 28
For the year Yc = a + bx
1992 = 88.43 + 5.04 (–3) = + 73.31
1993 = 88.43 + 5.04 (–2) = + 78.35
1990 12
1991 10
1992 14
1993 11
1994 13
1995 15
1996 16
Solution:
y xy
a= and b=
n x 2
91 21
a 13 b 0.75
7 28
Amity Directorate of Distance and Online Education
146 Quantitative Techniques in Management
2
n = 5 y = 266 x = 0 x = 10 xy = 48
y xy
Yc = a + bx a n and b 2
x
266 48
a ; b ; a = 53.2 and b = 4.8
5 10
Solution:
2
n = 7 y = 175 x = 0 x = 28 xy = 46
175 46
a = b =
7 28
= 25 = 1.64
For the year, Yc = a + bx
1996 = 25 + 1.64(-3) = 20.08
1997 = 25 + 1.64(-2) = 21.72
1998 = 25 + 1.64(-1) = 23.36
1999 = 25 + 1.64(0) = 25
2000 = 25 + 1.64(1) = 26.64
2001 = 25 + 1.64(2) = 28.28
2002 = 25 + 1.64(3) = 29.92
2006 is Yc = a + bx = 25 + 1.64(7) = 36.48
Notes I llustration - 7
Determine the trend values by the method of least squares from the following data Show the
actual time and trend line on graph.
2
N = 7 y = 420 x = 0 x = 28 xy = -146 Yc = 420
Where, Yc = a + bx
y 420 xy 146
a = = = 60 b = = = -5.21
N 7 x2 28
3.15 Summary
Correlation means that between two series or groups of data there exists some causal
connections.” At another place, he says, “If it is proved true that in a large number of instances
two variables tend always to fluctuate in the same or in opposite directions, we consider that the
fact is established and that a relationship exists. The relationship is called correlation.
Scatter Diagram is a graph of observed plotted points where each points represents the values
of X & Y as a co-ordinate. It portrays the relationship between these two variables graphically.
Probable error of Co-efficient of Correlation gives us the two limits within which the co-efficient
of correlation of Series selected at random from the same universe is likely to fall.
Notes Cyclical Analysis: Its refer to the long term swing about a trend line or curve.
These cycles, as they are sometimes called may or may not be periodic, i.e.,
they may or may not follow exactly similar pattens after equal intervals of time.
Random Analysis: Random or erratic variations are variations which occur in
a completely unpredictable fashion.
3.22 Bibliography
1. Srivastava V. K. etal – Quantitative Techniques for Managerial Decision Making,
Wiley Eastern Ltd
2. Richard, I.Levin and Charles A.Kirkpatrick – Quantitative Approaches to
Management, McGraw Hill, Kogakusha Ltd.
3. Prem.S.Mann, Introductory Statistics, 7th Edition, Wiley India, 2016.
4. Aczel A.D. and Sounderpandian J., “Complete Business Statistics”, 6th edition,
Tata McGraw Hill Publishing Company Ltd., New Delhi, 2012.
5. Anderson D.R., Sweeney D.J. and Williams T.A., Statistics for business and
economics, 11th edition, Thomson (South – Western) Asia, Singapore, 2012.
6. Budnik, Frank S Dennis Mcleaavey, Richard Mojena – Principles of Operation
Research - AIT BS New Delhi.
7. Sharma J K – Operation Research- theory and applications-Mc Millan,New Delhi
8. Kalavathy S. – Operation Research – Vikas Pub Co
9. Gould F J – Introduction to Management Science – Englewood Cliffs N J Prentice
Hall.
10. Naray J K, Operation Research, theory and applications – Mc Millan, New Dehi.
11. Taha Hamdy, Operations Research, Prentice Hall of India
12. Tulasian: Quantitative Techniques: Pearson Ed.
13. Vohr.N.D. Quantitative Techniques in Management, TMH.
11. Stevenson W.D, Introduction to Management Science, TMH.
Notes
Notes
Structure:
4.1 Introduction of Probability Theories
4.2 Concepts
4.3 Addition and Multiplication Theorems
4.4 Probability Distributions
4.5 Sampling
4.6 Steps in Developing a Sampling Plan
4.7 Sampling Techniques
4.8 Merits and Limitations of Sampling
4.9 Sampling Distribution
4.10 Hypothesis
4.11 Null Hypothesis and Alternative Hypothesis
4.12 t – test
4.13 F – test
4.14 Non–Parametric Tests (chi – square test)
4.15 Summary
4.16 Check Your Progress
4.17 Questions and Exercises
4.18 Key Terms
4.19 Check Your Progress: Answers
4.20 Case Study
4.21 Further Readings
4.22 Bibliography
Objectives
Definition of Probability
m
P(A) =
n
Where, ‘m’ is the number of favourable outcomes of an event ‘A’ and ‘n’ is the total
number of outcomes of the experiment.
Meaning of Probability
Probability is the likelihood that an event will occur and is calculated by dividing the
number of favourable outcomes by the total number of possible outcomes.
Example: There are only two possible outcomes when a coin is flipped, the result
is either heads or tails. And so the probability of getting heads is 1 out of 2 or ½ (50%).
Experiment:
Experiment refers to describe an act which can be repeated under some given
conditions. Random experiments are those experiments whose results depend on chance
such as tossing a coin, throwing a dice etc. The outcomes of random experiment are
events.
As previously stated, an experiment is a process that produces outcomes. Examples
of business oriented experiments with outcomes that can be statistically analyzed might
include the following:
1. Interviewing 20 randomly selected consumers and asking them which brand of
appliance they prefer.
2. Sampling every 200th bottle of ketchup from an assembly line and weighing
the contents.
3. Testing new pharmaceutical drugs on samples of cancer patients and measuring
the patient's improvement.
Random Experiment:
Sample Space:
The set of all possible outcomes of a random experiment is the sample space. The
sample space is denoted by S. The outcome of the random experiment are sample points
or outcomes or cases. For examples: In tossing of two coins the outcome are Head &
Tail
Hence Sample Space ‘S’ = {HH, HT, TT, TH}
Events:
Two events are said to be mutually exclusive or incompatible when both cannot
happen simultaneously in a single trial. i.e., the occurrence of any one of them precludes
the occurrence of the other. For example, if a single coin is tossed, either head or tail
can turn up. If we consider at one point of time, a person can be alive or dead.
Two or more events are said to be independent when the outcome of one does not
affect, and is not affected by the other. For example, if a coin is tossed twice, the result
of second would not be affected by the result of the first tossing.
Dependent events are those in which the occurrence or non-occurrence of one events
in other trails. For example, if a card is drawn from a pack of playing cards and is not
replaced, this will alter the probability that the second card is drawn.
Events are said to be equally likely when one does not occur more often than the
others. For example, if an unbiased dice is thrown, each face may be expected to be
observed approximately the same number of times in the long run.
In case of simple events, consider the probability of the happening or not happening
of single events. For example: If the are choosing two black coins from a box containing
10 white and 5 black coins.
In case of compound events, consider the joint occurrence of two or more events.
For example, if are interested in choosing or drawing 3 white coins in first draw and 2
black coins in second draw from a bag containing 10 white and 5 black coins.
Exhaustive Events:
Events are said to be exhaustive when their totality includes all the possible
outcomes of a random experiment. For example, while throwing a dice, the possible
outcomes are 1, 2, 3, 4, 5 and 6 and hence the exhaustive number of cases is 6.
Complementary Events:
Elementary Events:
Events that cannot be decomposed or broken down into other events are called
elementary events. Elementary events are denoted by lowercase letter.
Combinations of Events
Notes
Disjoint Events
Two events are called disjoint if they can not happen at the same time
Events A and B are disjoint means that the intersection of A and B is zero
Example: Coin is tossed twice
S = {HH, TH, HT, TT}
Events A = {HH} and B = {TT} are disjoint
Events A = {HH, HT} and B = {HH} are not disjoint
Probability Rule: If A and B are disjoint events then
P(A or B) = P(A) + P(B)
Additional Theorem
The addition theorem of probability states that if two events A and B are mutually
exclusive, the probability of occurrence of either A or B is sum of the individual probability
of A and B.
The rule of addition applies to the following situation. When there are two events,
the probability that either event occurs is to be known.
If A and B are two events associated with an experiment, then probability of
occurence of events A or B or both A and B is given by
P(A B) = P(A) + P(B) – P(A B)
Corollary 1:
n ( A B) 0
P(A B) = n ( S) = n( S ) = 0
Corollary 2:
n( S )
But, P(S) =
n( S ) = 1 . . . (ii)
P( A ) = 1 P(A)
Corollary 3:
Corollary 4:
(A B) bA Bg = S and (A B) bA Bg =
By Corollary 2,
b g
P A B = 1 P(A B) i.e., P(A B) = 1 P A B b g
This Corollary is very important, as it is useful in solving many examples. It can be
interpreted as
P(Occurence of atleast one event) 1 P (Occurence of none of the events)
Example:
If A and are independent events then prove that A and B are independent events.
Solution:
The rule of multiplication applies to the situation when the probability of the
intersection of two events is to be known that is, the probability that two events (Event
A and Event B) both occur.
If A and B are any two events associated with an experiment, then the probability
of simultaneous occurence of events A and B is given by
P(A B) = P(A) . P(B/A) = P(B) . P(A/B)
Corollary 1:
If A and B are independent events, then P(A/B) = P(A) and P(B/A) = P(B), hence
multiplication theorem in such a case will be
P(A B) = P(A) . P(B)
Note: If A, B and C are there independent events associated with an experiment,
then
P(A B C) = P(A) . P(B) . P(C)
In general, if A1, A2, . . . . An are n independent events, then probability of
simultaneous occurence of these n events is
P(A1 A2 . . . . An) = P(A1) P(A2) . . . P(An)
Notes Corollary 2:
2 2 3
If P(A) = , then P( A ) = 1 P(A) = 1 =
5 5 5
The odds in favour of A is given by
2 3
: i.e., 2 : 3
5 5
The odds against A is given by
3 2
: i.e., 3 : 2
5 5
Conversely, if odds in favour of A are a : b, then to find the probability of A, we divide
by (a + b).
a b
i.e., odds in favour of A can be written as
ab : ab
a b
P(A) = a b and P( A ) = a b
Illustration - 1
A ball is drawn at random from a box containing 6 red balls, 4 white balls and 5
blue balls. Find the probability that the ball drawn is not red.
Solution:
6
Probability of drawing a red ball =
15
6 9
Probability of not red = 1 =
15 15
Illustration - 2 Notes
In a single throw of two dice what is the probability of getting a total of 9?
Solution:
When two dice are thrown, the possible combinations are 62 = 36,
The cases which have a total of 9 are (3, 6), (4, 5), (5, 4), (6, 3) = 4
4 1
The probability = 36 9
Illustration - 3
A pair of unbiased dice is thrown. If the two numbers appearing be different, find
the probabilities that (i) the sum 6 (ii) the sum is 5 or less.
Solution:
If two dice are thrown there are 36 (= 6 × 6) exhaustive, mutually and equally likely
outcomes (i) favourable cases are (1, 5), (2, 4), (4, 2), (5, 1) i.e. only 4 cases.
4 1
P(total six in sum) = =
36 9
8 2
Required probability = =
36 9
Illustration - 4
If 8 persons are arranged at random (i), in a line, (ii) in a ring find the probability
that two particular person will be always together.
Solution:
(i) 8 persons can be arranged in a line 8! ways, which are mutually exclusive,
exhaustive and equally likely.
Now taking 2 particular person as one, 7 = (1 + 6) person can be arranged in 7!
Ways. Again these persons can be arranged almost themselves in 2! Ways. So, the total
number of favorable cases 7! 2!
7! 2! 1
Required probability = =
8! 4
Notes (ii) 8 persons can be arranged in a ring in 7! Ways. As before number of favourable
cases is 6! 2!
6! 2! 2
Required probability = =
7! 7
Illustration - 5
A company wants to select a team leader from among the following candidates:
a) Male, age 40
b) Male, age 43
c) Female, age 38
d) Female, age 29
e) Male, age 39
What is the probability that the team leader selected, will be
1) Either male or aged above 41 years.
2) Either female or aged below 35 years.
Solution:
3 1 1 4 1 3
= = =
5 5 5 5 5 5
(2) P(female) below 35 years
2 1 1 3 1 2
= =
5 5 5 5 5 5
Illustration - 6
What is probability of getting 3 white balls in a draw of 3 balls from a box containing
5 white and 4 black balls?
Solution:
987
3 balls can be drawn from 9 balls in 9C3 =
3 2 1 = 84 ways
Total number of possible cases = 84.
Total number of white balls in the box = 5.
543
3 white balls can be drawn from 5 white balls in 5C3 =
3 2 1 = 10 ways
Number of cases favourable to the event of getting 3 white balls = 10. Notes
10 5
Probability of drawing 3 white balls = =
84 42
Illustration - 7
Solution:
1st part: 3 men can be selected out of 20 men in 20C3 ways, so total number of
possible cases is 20C3. Again 3 men to be graduates, these 3 men are to be selected
out of 5 graduates, which can be done in 5C3. Now the total number of favourable cases
is 5C3.
5
C3 10 1
Required probability = 20 = =
C3 1140 114
5! 3! 5. 4 .3
Calculation: 5C3 = =
2! 3. 2 = 10
20.19.18
20C = = 1140
3 3. 2
2nd Part: At least one graduate means it may be 1, 2 or 3 graduates. now it will
be easier to find the complementary event i.e. none is graduate (or all 3 are non-graduate).
So 3 men (non-graduate) can be selected out 15 ( = 10 – 5) non-graduates in 15C3 ways.
Total number of favourable cases is 15C3.
15
C3 455 91 137
Required probability = 1 20 =1 =1 =
C3 1140 228 228
15.14 .13
Calculation: 15C = = 455,
3 3. 2.1
20.19.18
15C =
3 3. 2.1 = 1140
Illustration - 8
An urn contains 8 red and 3 blue marbles. If 2 marbles are drawn at random, find
the probability that (i) Both are red and (ii) One of them is red and another is blue.
Solution:
Total number of marbles in the urn = 8 + 3 = 11
11 10
11
Two marbles can be drawn from 11 in C2 ways =
2 1 = 55 ways
87
2 marbles can be drawn from 8 red marbles in 8C2 ways =
2 1 = 28 ways
28
The probability that the two marbles drawn at random are red = = 0.509
55
(ii) Probability that one of them is red and another is blue:
One red marble can be drawn from 8 red marbles in 8c1 ways and 1 blue marble can
be drawn from 3 blue marbles in 3c1 ways.
The number of favourable cases is 8c1 x 3c1 and the required probability is
8c1 x 3c1 8 x3 24
P = = = = 0.436
11c 2 55 55
Illustration - 9
Three cards are drawn at random from a pack of 52 cards. Find the probability that:
(i) They are king, Queen and Knave
(ii) All of them are Aces
(iii) All of them are Spades
(iv) Two are red and one Black
(v) They are not face cards.
Solution:
There cards can be drawn at random from a well shuffled pack of 52 cards in 52C3
ways.
(i) One King can be drawn out of 4 Kings (in a full pack) in 4C3 ways. Similarly one
Queen in 4C3 and one Knave in 4C3 ways. Since each way of drawing a King can be
associate with each of way of drawing Queen and Knave, so number of favourable cases
is 4C1 × 4C1 × 4C1
4
C1 4 C1 4 C1 444 16
Required probability = 52 = =
C3 22100 5525
(ii) There Aces can be drawn out of four Aces in 4C3 ways
4
C3
Required probability = 52
C3
(iii) 3 Spades can be drawn out of 13 Spades in 13C3 ways.
13
C3
Required probability = 52
C3
26 26
C2 C1
(iv) Required probability = 52
C3
(v) Number of cards other than face = 52 – 12 = 40, as there are 12 face cards. Notes
40
C3
Required probability = 52
C3
Illustration - 10
Solution:
There are 4 Jacks and 13 Spades in a pack of 52 cards. Also Spade has one Jack in
it.
13
Probability of getting a Spade =
52
4
Probability of getting a Jack =
52
1
Probability of getting a Jack Spade =
52
Let, A be the event of getting a Spade
B be the event of getting a Jack.
As there is a Jack in Spade, the events are not mutually exclusive events. Hence,
the events are not mutually exclusive events. Hence, the probability of getting a Jack or a
Spade is
P(A B) = P(A) + P(B) – P(A B)
13 4 1 16 4
=
52 52 52 52 13
Notes Definitions
1. Two events are mutually exclusive or disjoint if they cannot occur at the same
time.
2. The probability that Event A occurs, given that Event B has occurred, is called
a conditional probability. The conditional probability of Event A, given Event B,
is denoted by the symbol P (A|B).
3. The complement of an event is the event not occurring. The probability that Event
A will not occur is denoted by P (A').
4. The probability that Events A and B both occur is the probability of the
intersection of A and B. The probability of the intersection of Events A and B
is denoted by P (A B). If Events A and B are mutually exclusive, P (A B)
= 0.
5. The probability that Events A or B occur is the probability of the union of A
and B. The probability of the union of Events A and B is denoted by P(A
B).
6. If the occurrence of Event A changes the probability of Event B, then Events
A and B are dependent. On the other hand, if the occurrence of Event A does
not change the probability of Event B, then Events A and B are independent.
Continuous random variables are those that take on any value including fractions
and decimals. Continuous random variables give rise to continuous probability
distributions. Continuous is the opposite of discrete.
If a random variable is a continuous variable, its probability distribution is called a
continuous probability distribution.
A continuous random variable is a random variable that has an infinite number of
possible values that is not countable .It is usually the result of some type of measurement.
Example:
1. The income in a year for a family.
2. The distance students travel to class.
3. The time it takes an executive to drive to work.
4. The length of an afternoon nap.
5. The length of time of a particular phone call.
A continuous probability distribution differs from a discrete probability distribution in
several ways.
1. The probability that a continuous random variable will assume a particular value
is zero.
2. As a result, a continuous probability distribution cannot be expressed in tabular
form.
3. Instead, an equation or formula is used to describe a continuous probability
distribution.
Most often, the equation used to describe a continuous probability distribution is
called a probability density function. Sometimes, it is referred to as a density function,
a PDF or a pdf. For a continuous probability distribution, the density function has the
following properties:
(i) Since the continuous random variable is defined over a continuous range of
values i.e., called the domain of the variable, the graph of the density function
will also be continuous over that range.
(ii) The area bounded by the curve of the density function and the x-axis is equal
to 1, when computed over the domain of the variable.
(iii) The probability that a random variable assumes a value between a and b is equal
to the area under the density function bounded by a and b.
Notes according to the probability of that event occurring. The expected value of X is usually
written as E(X) or m.
E(X) = S P(X = x)
The expected value is the sum of: [(each of the possible outcomes) × (the probability
of the outcome occurring)].
In more concrete terms, the expectation is what you would expect the outcome of
an experiment to be on average.
Example:
What is the expected value when we roll a fair die?
There are six possible outcomes: 1, 2, 3, 4, 5, 6. Each of these has a probability
of 1/6 of occurring. Let X represent the outcome of the experiment.
Therefore P(X = 1) = 1/6 (this means that the probability that the outcome of the
experiment is 1 is 1/6)
P(X = 2) = 1/6 (the probability that throw a 2 is 1/6)
P(X = 3) = 1/6 (the probability that throw a 3 is 1/6)
P(X = 4) = 1/6 (the probability that throw a 4 is 1/6)
P(X = 5) = 1/6 (the probability that throw a 5 is 1/6)
P(X = 6) = 1/6 (the probability that throw a 6 is 1/6)
E(X) = 1 × P(X = 1) + 2 × P(X = 2) + 3 × P(X = 3) + 4 × P(X = 4) + 5 × P(X =
5) + 6 × P(X = 6)
Therefore, E(X) = 1/6 + 2/6 + 3/6 + 4/6 + 5/6 + 6/6 = 7/2
The Mean of a Probability Distribution
The mean is a typical value used to represent the central location of a probability
distribution.
The mean of a probability distribution is also referred to as its expected value.
The Variance and Standard Deviation of a Probability Distribution
The computational steps are:
1. Subtract the mean from each value, and square this difference.
2. Multiply each squared difference by its probability.
3. Sum the resulting products to arrive at the variance.
Binomial Distribution
If a group of patients is given a new drug for the relief of a particular condition, then
the proportion p being successively treated can be regarded as estimating the population
treatment success rate .
The sample proportion p is analogous to the sample mean x , in that if we score
zero for those s patients who fail on treatment and unity for those r who succeed, then
p = r/n, where n = z + s is the total number of patients treated. Thus p also represents
a mean.
Data which can take only a 0 or 1 response, such as treatment failure or treatment Notes
success, follow the binomial distribution provided the underlying population response rate
does not change. The binomial probabilities are calculated from
Probability of obtaining exactly ‘r’ successes in a given number of ‘n’ Bernoulli trials
is
f(r) = P(X = r) = n Cr pr qn - 1
For r = 1, 2, ….. n
Where,
p = probability of success on a single trial.
q = probability of failure on a single trial.
e = number of (Bernoulli) trials.
r = number of successes in ‘n’ trials.
Mean Binomial distribution ( ) = np
Variance of Binomial distribution( 2) = npq
For successive values of R from 0 through to n. In the above n! is read as n factorial
and R! as R factorial. For R = 4, R! = 4 × 3 × 2 × 1 = 24. Both 0! and 1! are taken
as equal to unity. The shaded area marked in Figure corresponds to the above expression
for the binomial distribution calculated for each of R = 8, 9, ..., 20 and then added. This
area totals 0.1018. So the probability of eight or more responses out of 20 is 0.1018.
For a fixed sample size n the shape of the binomial distribution depends only on
. Suppose n = 20 patients are to be treated and it is known that on average a quarter
or = 0.25 will respond o this particular treatment. The number of responses actually
observed can only take integer values between 0 (no responses) and 20 (all respond).
The binomial distribution for this case is illustrated in below figure.
The distribution is not symmetric, it has a maximum at five responses and the height
of the blocks corresponds to the probability of obtaining the particular number of responses
from the 20 patients yet to be treated. It should be noted that the expected value for r,
the number of successes yet to be observed if we treated n patients, is n . The potential
variation about this expectation is expressed by the corresponding standard deviation
SE(r) = [n (1- )]
Figure shows the Normal distribution arranged to have µ = n = 5 and s = v [n (1-
)] = 1.94, superimposed on to a binomial distribution with p = 0.25 and n = 20. The
Normal distribution describes fairly precisely the binomial distribution in this case.
If n is small, however or close to 0 or 1, the disparity between the Normal and
binomial distributions with the same mean and standard deviation, similar to those
illustrated in Figure, increases and the normal distribution can no longer be used to
approximate the binomial distribution. In such cases the probabilities generated by the
binomial distribution itself must be used.
It is also only in situations in which reasonable agreement exists between the
distributions use the confidence interval expression given previously. For technical
reasons, the expression given for a confidence interval for is an approximation. The
approximation will usually be quite good provided p is not too close to 0 or 1, situations
in which either almost none or nearly all of the patients respond to treatment. The
approximation improves with increasing sample size n.
Amity Directorate of Distance and Online Education
172 Quantitative Techniques in Management
Notes
Fig: Binomial distribution for n = 20 with = 0.25 and the Normal approximation
Illustration - 1
6
Probability of a delinquent account P = = 0.3
20
Probability of a non-delinquent account q = 1 – p = 1 – 0.3 = 0.7
(i) Probability that auditor finds (r = 2) delinquent accounts in a random choice of(n = 5)
accounts is given by Binomial distribution as:
p(r) = n Cr pr qn - 1
Illustration - 2
Solution: Notes
Probability of success p = 0.60
( P = 1 – q and q = 0.4)
Probability of failure q = 0.40 (given)
Number of trails = 6
By Binomial distribution, the probability that ‘r’ candidates the examination is
p(r) = n Cr pr qn - 1
= 6 n Cr pr qn - 1
Illustration - 3
The probability that an entering student will graduate is 0.4. Determine the probability
that out of 5 students at least one will graduate.
Solution:
Illustration - 4
Find the mean and variance of Binomial distribution with parameters n and p.
Solution:
n n
= n. nCx . px qn . q
x
x0
n
FG pIJ x
= qn
x0
x. nCx
H qK
L F pI F pI F pI F pI OP
= M 0 + 1 . nC G J + 2 . nC G J + 3 . nC G J +..+ n.nC G J
2 3 n
MN
qn
HqK
1
H qK 2 H q K HqK 3 n
PQ
LMn FG pIJ 2. n ( n 1) FG pIJ 3. n ( n 1) ( n 2) FG pIJ ... nFG pIJ OP
2 3 n
=
N HqK 2 H qK 3! H qK H q K PQ
F pI L
= q . n G J M 1 + (n 1) G J +
F pI ( n 1) ( n 2) FG pIJ + ... + FG pIJ 2 n 1
OP
n
H q K MN HqK 2! H qK HqK PQ
F pI L p O = q np LM( p q ) OP = q . np(1)
= q . n G J M1 P
n 1 n 1 n
n
HqK N q Q q N q
n
Q q .q n 1 n 1
= np
n
E(X2) = x2p(x)
x0
= (x2 x + x) p(x)
x0
n n
n n
=
x
x(x 1) nCx px qn
x0
n
FG pIJ x
=
x0
x(x 1) nCx
H qK qn
LM F pI 2
FG pIJ 3
FG pIJ 4
FG pIJ n OP Notes
0 + 0 + 2.1nC G J
= qn
MN H qK 2 + 3.2nC3
H qK + 4.3nC4
H qK + ... + nCn
HqK PQ
LM FG IJ + 3.2 . ( n 1) ( n 2) FG pIJ + ... + n(n 1) FG pIJ
n ( n 1) p
2 3 n OP
= qn
MN 2.1
2 H K
q 3.21. H qK HqK PQ
F pI L F pI ( n 2) ( n 3) FG pIJ + .. + FG pIJ
. n(n - 1) G J M 1 + (n - 2) G J +
2 2 n2 OP
= qn
H q K MN HqK 2 H qK H qK PQ
F pI L p O = q . n(n 1) p . ( p q )
. n(n 1) G J M1 P
2 n2 2 n2
= qn
H qK N qQ q n
q 2 n2
Poisson Distribution
The Poisson distribution is used to describe discrete quantitative data such as counts
in which the populations size n is large, the probability of an individual event is small,
but the expected number of events, n is moderate say five or more.
Typical examples are the number of deaths in a town from a particular disease per
day or the number of admissions to a particular hospital.
Poisson distribution is a discrete probability distribution developed by French
Mathematician, S. Poisson. Poisson distribution is used when the chance of any individual
event being a success is small. The distribution is given by the following formula.
e m mr
P(r) n =
r!
Where r = 0, 1, 2, 3, 4, ………
e = 2.7183 (the base of natural logarithms)
m = mean of poisson distribution = np
(average no. of occurrences of an event)
The poisson distribution is a discrete distribution with a single parameter ‘m’. As
‘m’ increases, the distribution shifts to the right. All Poisson probability distribution are
skewed to right. Poisson distribution is a distribution of rate events (the probabilities tend
to be high for small numbers of occurrences). Mean = variance = m (for a Poisson
distribution).
Notes Illustration - 5
Suppose that on an average 1 house in 1000 in a certain district has a fire during
a year. If there are 3000 houses in that districts, What is the probability that exactly 4
houses will have a fire during the year?
Solution:
1
= = 0.001
1000
Since n is large and p is small, Poisson distribution is applied.
Mean or expected number of houses to each fire is
m = np = 3,000 × 0.001 = 1
Probability that ‘r’ occurrences will take place in ‘n’ is
e m mr
p(r) =
r!
Probability that exactly 4 houses will have a fire during the year is
e 3 m4 0.050 81
p(4) = = = 0.169
4! 24
Illustration - 6
In a Poisson distribution, the probability P(x) for X = 0 is 10 percent. Find the mean
of the distribution.
Solution:
By Poisson distribution,
e m mr
P(r) =
r!
where, m mean of the distribution
Given P(0) = 0.10 (i.e. 10%)
e m 1 e m 1
P(0 = = 0.1 = 0.1 e- m = 0.1
1 1
m = 2.3026
(To obtain ‘m’, take natural log of 0.1)
Mean of the distribution = 2.3026.
1
P(x) =
2
Where, x = Values of continuous random variable
= Mean of the normal random variable
E = Mathematical constant = 2.7183
= Mathematical constant = 3.1416
= Standard deviation of normal random.
Notes shaped’ distribution known as a Normal distribution. One such example is the histogram
of the birth weight in kilograms of the 3,226 new born babies.
Illustration - 1
1000 light bulbs with a mean life of 120 days are installed in a firm. Their length
of life is normally distributed with a standard deviation of 20 days. How many bulbs will
expire in less than 90 days?
Solution:
Notes x = 90 x = 120
z = 1.5 z = 0
Area between Z = 0 and Z = -1.5 is 0.4332
Area less than 90 days = 0.5 - 0.4332 = 0.668
Therefore the number of bulbs that will expire in loess than 90 days is 0.0668 x 1000
= 66.8 bulbs.
Illustration - 2
X 15
Then, Z is N (0,1)
3
x 15 13 15
a) P(x > 13) = P = P(Z> – 0.666)
3 3
-0.666
= area from –0.666 to = 0.7454
x 15 19 15
b) P(x < 19) = P = P(Z > 1.33)
3 3
= (area from – (0 ) – (area from 1.33 to )
= 1 – 0.1515 = 0.8485
Hypergeometric Distribution
K N K
k n k
P ( X k)
N
n
6 14
4 1
P ( X k )
20
5
Solution:
14
C6
4
14C1 20
C5 15
15504
= 0.0135
The binomial distribution doesn’t apply here, because the cards are not replaced once
they are drawn. In other words, the trials are not independent events. For example, for
1 red card, the probability is 6/20 on the first draw. If that card is red, the probability of
choosing another red card falls to 5/19.
When given x, N, n and k, we can compute the hypergeometric probability based on
the following formula:
The population consists of N items, k of which are successes. And a random sample
drawn from that population consists of nitems, x of which are successes. Then the
hypergeometric probability is:
h(x; N, n, k) = [ kCx ] [ N-kCn-x ] / [ NCn ]
Exponential Distribution
Another useful continuous distribution is the exponential distribution, which has the
following probability density function:
f(x) = λe λx For, x 0.
For, λ = 0.5, 1 and 2, the shapes of the exponential density curve are:
Observe that the greater the rate, the faster the curve drops. Or, the lower the rate,
the flatter the curve.
Several useful formulas are:
-λx
P{X x} = 1 - e
4.5 Sampling
Sampling is an important concept which is practiced in every activity. Sampling
involves selecting a relatively small number of elements from a large defined group of
elements and expecting that the information gathered from the small group will allow
judgments to be made about the large group. The basic idea of sampling is that by
selecting some of the elements in a population, the conclusion about the entire population
is drawn. Sampling is used when conducting census is impossible or unreasonable. In
a census method a researcher collects primary data from every member of a defined target Notes
population. It is not always possible or necessary to collect data from every unit of the
population. The researcher can resort to sample survey to find answers to the research
questions. However, they can do more harm than good if the data is not collected from
the people, events or objects that can provide correct answers to the problem. The process
of selecting the right individuals, objects or events for the purpose of the study is known
as sampling
Meaning of Sampling
Purpose of Sampling
There are several reasons for sampling. They are explained below:
(i) Lower cost: The cost of conducting a study based on a sample is much lesser
than the cost of conducting the census study.
(ii) Greater accuracy of results: It is generally argued that the quality of a study
is often better with sampling data than with a census. Research findings also
substantiate this opinion.
(iii) Greater speed of data collection: Speed of execution of data collection is
higher with the sample. It also reduces the time between the recognition of a
need for information and the availability of that information.
(iv) Availability of population element: Some situations require sampling. When
the breaking strength of materials is to be tested, it has to be destroyed. A
census method cannot be resorted as it would mean complete destruction of
all materials. Sampling is the only process possible if the population is infinite.
The sampling technique has the following good features of value and significance:
1. Economy: Sampling technique brings about cost control of a research project
as it requires much less physical resources as well as time than the census
technique.
2. Reliability: In sampling technique, if due diligence is exercised in the choice
of sample unit and if the research topic is homogenous then the sample survey
can have almost the same reliability as that of census survey.
3. Detailed Study: An intensive and detailed study of sample units can be done
since their number is fairly small. Also multiple approaches can be applied to
a sample for an intensive analysis.
4. Scientific Base: As mentioned earlier this technique is of scientific nature as
the underlined theory is based on principle of statistics.
5. Greater Suitability in most Situations: It has a wide applicability in most
situations as the examination of few sample units normally suffices.
6. Accuracy: The accuracy is determined by the extent to which bias is eliminated
from the sampling. When the sample elements are drawn properly some sample
elements underestimates the population values being studied and others
overestimate them.
In order to reach a clear conclusion, the sampling should possess the following
essentials:
1. It must be representative: The sample selected should possess the similar
characteristics of the original universe from which it has been drawn.
2. Homogeneity: Selected samples from the universe should have similar nature
and should not have any difference when compared with the universe.
3. Adequate Samples: In order to have a more reliable and representative result,
a good number of items are to be included in the sample.
4. Optimization: All efforts should be made to get maximum results both in terms
of cost as well as efficiency. If the size of the sample is larger, there is better
efficiency and at the same time the cost is more. A proper size of sample is
maintained in order to have optimized results in terms of cost and efficiency.
resources, time frame, advanced knowledge of the target population, scope of the research Notes
and perceived statistical analysis needs.
5) Determine necessary sample sizes and overall contact rates
The sample size is decided based on the precision required from the sample
estimates, time and money available to collect the required data. While determining the
sample size due consideration should be given to the variability of the population
characteristic under investigation, the level of confidence desired in the estimates and the
degree of the precision desired in estimating the population characteristic. The number
of prospective units to be contacted to ensure that the estimated sample size is obtained
and the additional cost involved should be considered. The researcher should calculate
the reachable rates, overall incidence rate and expected completion rates associated with
the sampling situation.
6) Creating an operating plan for selecting sampling units
The actual procedure to be used in contacting each of the prospective respondents
selected to form the sample should be clearly laid out. The instruction should be clearly
written so that interviewers know what exactly should be done and the procedure to be
followed in case of problems encountered, in contacting the prospective respondents.
7) Executing the operational plan
The sample respondents are met and actual data collection activities are executed
in this stage. Consistency and control should be maintained at this stage.
I. Probability Sampling
Probability sampling is where each sampling unit in the defined target population
has a known non-zero probability of being selected in the sample. The actual probability
of selection for each sampling unit may or may not be equal depending on the type of
probability sampling design used. Specific rules for selecting members from the operational
population are made to ensure unbiased selection of the sampling units and proper sample
representation of the defined target population. The results obtained by using probability
sampling designs can be generalized to the target population within a specified margin
of error.
Probability samples are characterised by the fact that, the sampling units are
selected by chance. In such a case, each member of the population has a known, non-
zero probability of being selected. However, it may not be true that all samples would
have the same probability of selection, but it is possible to say the probability of selecting
any particular sample of a given size. It is possible that one can calculate the probability
that any given population element would be included in the sample. This requires a precise
definition of the target population as well as the sampling frame. Probability sampling
Amity Directorate of Distance and Online Education
186 Quantitative Techniques in Management
Notes techniques differ in terms of sampling efficiency which is a concept that refers to trade
off between sampling cost and precision. Precision refers to the level of uncertainty about
the characteristics being measured. Precision is inversely related to sampling errors but
directly related to cost. The greater the precision, the greater the cost and there should
be a trade-off between sampling cost and precision. The researcher is required to design
the most efficient sampling design in order to increase the efficiency of the sampling.
The different types of probability sampling designs are discussed below:
1) Simple Random Sampling
The systematic random sampling design is similar to simple random sampling but
requires that the defined target population should be selected in some way. It involves
drawing every nth element in the population starting with a randomly chosen element
between 1 and n. In other words individual sampling units are selected according their
position using a skip interval. The skip interval is determined by dividing the sample size
into population size. For example, if the researcher wants a sample of 100 to be drawn Notes
from a defined target population of 1000, the skip interval would be 10(1000/100). Once
the skip interval is calculated, the researcher would randomly select a starting point and
take every 10th until the entire target population is proceeded through. The steps to be
followed in a systematic sampling method are enumerated below:
(i) Total number of elements in the population should be identified
(ii) The sampling ratio is to be calculated ( n = total population size divided by
size of the desired sample)
(iii) A sample can be drawn by choosing every nth entry
Two important considerations in using the systematic random sampling are:
(i) It is important that the natural order of the defined target population list be
unrelated to the characteristic being studied.
(ii) Skip interval should not correspond to the systematic change in the target
population.
Advantages and Disadvantages
The major advantage is its simplicity and flexibility. In case of systematic sampling
there is no need to number the entries in a large personnel file before drawing a sample.
The availability of lists and shorter time required to draw a sample compared to random
sampling makes systematic sampling an attractive, economical method for researchers.
The greatest weakness of systematic random sampling is the potential for the hidden
patterns in the data that are not found by the researcher. This could result in a sample
not truly representative of the target population. Another difficulty is that the researcher
must know exactly how many sampling units make up the defined target population. In
situations where the target population is extremely large or unknown, identifying the true
number of units is difficult and the estimates may not be accurate.
Stratified random sampling requires the separation of defined target population into
different groups called strata and the selection of sample from each stratum. Stratified
random sampling is very useful when the divisions of target population are skewed or when
extremes are present in the probability distribution of the target population elements of
interest. The goal in stratification is to minimize the variability within each stratum and
maximize the difference between strata. The ideal stratification would be based on the
primary variable under study. Researchers often have several important variables about
which they want to draw conclusions. A reasonable approach is to identify some basis
for stratification that correlates well with other major variables. It might be a single variable
like age, income etc. or a compound variable like on the basis of income and gender.
Stratification leads to segmenting the population into smaller, more homogeneous sets
of elements. In order to ensure that the sample maintains the required precision in terms
of representing the total population, representative samples must be drawn from each of
the smaller population groups.
There are three reasons as to why a researcher chooses a stratified random sample:
(i) To increase the sample’s statistical efficiency
(ii) To provide adequate data for analyzing various sub populations
(iii) To enable different research methods and procedures to be used in different
strata.
4) Cluster Sampling
Cluster sampling is a probability sampling method in which the sampling units are
divided into mutually exclusive and collectively exhaustive subpopulation called clusters.
Each cluster is assumed to be the representative of the heterogeneity of the target
population. Groups of elements that would have heterogeneity among the members within
each group are chosen for study in cluster sampling. Several groups with intragroup
heterogeneity and intergroup homogeneity are found. A random sampling of the clusters
or groups is done and information is gathered from each of the members in the randomly
chosen clusters. Cluster sampling offers more of heterogeneity within groups and more
homogeneity among the groups.
Single Stage and Multistage Cluster Sampling
In single stage cluster sampling, the population is divided into convenient clusters
and required number of clusters are randomly chosen as sample subjects. Each element
in each of the randomly chosen cluster is investigated in the study. Cluster sampling can Notes
also be done in several stages which is known as multistage cluster sampling. For
example: To study the banking behaviour of customers in a national survey, cluster
sampling can be used to select the urban, semi-urban and rural geographical locations
of the study. At the next stage, particular areas in each of the location would be chosen.
At the third stage, the banks within each area would be chosen. Thus multi-stage sampling
involves a probability sampling of the primary sampling units; from each of the primary
units, a probability sampling of the secondary sampling units is drawn; a third level of
probability sampling is done from each of these secondary units, and so on until the final
stage of breakdown for the sample units are arrived at, where every member of the unit
will be a sample.
Area Sampling
Area sampling is a form of cluster sampling in which the clusters are formed by
geographic designations. For example, state, district, city, town etc., Area sampling is
a form of cluster sampling in which any geographic unit with identifiable boundaries can
be used. Area sampling is less expensive than most other probability designs and is not
dependent on population frame. A city map showing blocks of the city would be adequate
information to allow a researcher to take a sample of the blocks and obtain data from
the residents therein.
Advantages and Disadvantages of Cluster Sampling
The cluster sampling method is widely used due to its overall cost-effectiveness and
feasibility of implementation. In many situations the only reliable sampling unit frame
available to researchers and representative of the defined target population, is one that
describes and lists clusters. The list of geographical regions, telephone exchanges, or
blocks of residential dwelling can normally be easily compiled than the list of all the
individual sampling units making up the target population. Clustering method is a cost
efficient way of sampling and collecting raw data from a defined target population.
One major drawback of clustering method is the tendency of the cluster to be
homogeneous. The greater the homogeneity of the cluster, the less precise will be the
sample estimate in representing the target population parameters. The conditions of intra-
cluster heterogeneity and inter-cluster homogeneity are often not met. For these reasons
this method is not practiced often.
5) Sequential/Multiphase Sampling
This is also called Double Sampling. Double sampling is opted when further
information is needed from a subset of groups from which some information has already
been collected for the same study. It is called as double sampling because initially a
sample is used in the study to collect some preliminary information of interest and later
a sub-sample of this primary sample is used to examine the matter in more detail The
process includes collecting data from a sample using a previously defined technique.
Based on this information, a sub sample is selected for further study. It is more convenient
and economical to collect some information by sampling and then use this information
as the basis for selecting a sub sample for further study.
6) Sampling with Probability Proportional to Size
When the case of cluster sampling units does not have exactly or approximately
the same number of elements, it is better for the researcher to adopt a random selection
process, where the probability of inclusion of each cluster in the sample tends to be
proportional to the size of the cluster. For this, the number of elements in each cluster
has to be listed, irrespective of the method used for ordering it. Then the researcher should
Notes systematically pick the required number of elements from the cumulative totals. The actual
numbers thus chosen would not however reflect the individual elements, but would indicate
as to which cluster and how many from them are to be chosen by using simple random
sampling or systematic sampling. The outcome of such sampling is equivalent to that
of simple random sample. This method is also less cumbersome and is also relatively
less expensive.
In non probability sampling method, the elements in the population do not have any
probabilities attached to being chosen as sample subjects. This means that the findings
of the study cannot be generalized to the population. However, at times the researcher
may be less concerned about generalizability and the purpose may be just to obtain some
preliminary information in a quick and inexpensive way. Sometimes when the population
size is unknown, then non probability sampling would be the only way to obtain data.
Some non-probability sampling techniques may be more dependable than others and could
often lead to important information with regard to the population.
Non-probability sampling does not involve random selection. It involves personal
judgement of the researcher rather than chance to select sample elements. Sometimes
this judgement is imposed by the researcher, while in other cases the selection of
population elements to be included is left to the individual field workers. The decision maker
may also contribute to including a particular individual in the sampling frame. Evidently,
non probability sampling does not include elements selected probabilistically and hence,
leaves a degree of sampling error associated with the sample. Sampling error is the degree
to which a sample might differ from the population. Therefore, while inferring to the
population, results could not be reported plus or minus the sampling error. In non-
probability sampling, the degree to which the sample differs from the population remains
Unknown However, we cannot come to a conclusion that sampling error is an inherent
of non probability sample. on-probability samples also yield good estimates of the
population characteristics. Since, inclusion of the elements in the sample are not
determined in a probabilistic way, the estimates obtained are not statistically projectable
to the population.
1) Convenience Sampling
samples in the development phases of constructs and scale measurements can have a Notes
serious negative impact on the overall reliability and validity of those measures and
instruments used to collect raw data. Another major drawback is that the raw data and
results are not generalizable to the defined target population with any measure of precision.
It is not possible to measure the representativeness of the sample, because sampling
error estimates cannot be accurately determined.
2) Judgment Sampling
3) Quota Sampling
Merits of Sampling
1. It saves time: Sampling method of data collection saves time because fewer
items are collected and processed. When the results are urgently required, this
method is very helpful.
2. It reduces cost: Since only a few and selected items are studied in sampling,
there is reduction in cost of money and reduction in terms of man hours.
3. More reliable results can be obtained: Through sampling, more reliable results
can be obtained because (a) there are fewer chances of sampling statistical
errors. If there is sampling error, it is possible to estimate and control the results.
(b) Highly experienced and trained persons can be employed for scientific
processing and analyzing of relatively limited data and they can use their high
technical knowledge and get more accurate and reliable results.
4. It provides more detailed information: As it saves time, money and labor,
more detailed information can be collected in a sample survey.
5. Dependency on Sampling method: Some times it so happens that one has
to depend upon sampling method alone because if the population under study
is finite, sampling method is the only method to be used. For example, if
someone’s blood has to be examined, it will become fatal to take all the blood
out from the body and study depending upon the total enumeration method.
6. Administrative convenience: The organization and administration of sample
survey are easy and convenient for the researcher.
7. More scientific: Since the methods used to collect data are based on scientific Notes
theory and results obtained can be tested, sampling is a more scientific method
of collecting data.
Limitations of Sampling
1. Less Accuracy: In comparison to census technique the conclusions derived
from samples are more liable to error. Therefore, sampling technique is less
accurate than the census technique.
2. Changeability of Units: If the units in the field of survey are liable to change
or if these are not homogenous the sampling technique will be very hazardous.
It is not scientific to extend the conclusions derived from one set of sample
to other sets which are dissimilar or are changeable.
3. Misleading Conclusions: If due care is not taken in the selection of samples
or if they are arbitrarily selected, the conclusions derived from them will become
misleading if extended to all units. For example, in assessing the monthly
expenditure of university students we select a sample study of only rich
students, our results will be highly erroneous if extended to all students.
4. Need for Specialized Knowledge: The sample technique can be successful
only if a competent and able researcher makes the selection. If this is done
by an average researcher, the selection is liable to be wrong.
5. Situational Impact: Under certain circumstances it is very difficult to use the
sampling technique. If the time is very short and it is not possible to make
selection of the sample, the technique cannot be used. Besides, if we need
100% accuracy the sampling technique cannot be used. It can also not be used
if the material is of heterogeneous nature.
6. Sample not representative: To make the sample representative is a difficult
task. If a representative sample is taken from the universe, the result is
applicable to the whole population. If the sample is not representative of the
universe the result may be false and misleading.
7. Complicated Process: Sometimes the sampling plan may be complicated and
requires more money, labor and time than a census method.
8. Personal bias: There may be personal biases and prejudices with regard to
the choice of technique and drawing of sampling units on the part of the
researcher.
This hypothesis is then tested with available evidence and a decision is made whether Notes
to accept this hypothesis or reject it. If this hypothesis is rejected, then we accept the
alternate hypothesis. This hypothesis is written as H 1 .
For testing hypothesis or test of significance we use both parametric tests and
nonparametric or distribution free tests. Parametric tests assume within properties of the
population, from which we draw samples. Such assumptions may be about population
parameters, sample size etc. In case of non-parametric tests, we do not make such
assumptions. Here we assume only nominal or ordinal data.
Characteristics of Hypothesis
Formulation of Hypothesis
Null hypothesis, H 0 : µ = 50
Alternative hypothesis H : µ = 50
1
or µ > 50
µ < 50
2. Establish a level of significance
The level of significance signifies the probability of committing Type 1 error á and
is generally taken as equal to 0.05. Sometimes, the value a is established as 0.01, but
it is at the discretion of the investigator to select its value, depending upon the sensitivity
of the study. To illustrate per cent level of significance indicates that a researcher is willing Notes
to take 5 per cent risk of rejecting the Null Hypothesis when it happens to be true.
3. Choosing a suitable test statistic
Now the researcher would choose amongst the various tests (i.e. z, t, 2 and f-
tests). Actually, for the purpose of rejecting or accepting the null hypothesis, a suitable
statistics called ‘test statistics’ is chosen. This means that H 0 is assumed to be really
true. Obviously due to sampling fluctuations, the observed value of the statistic based
on random sample will differ from the expected value. If the difference is large enough,
one suspects the validity of the assumption and rejects the null hypothesis ( H 0 ). On
the other hand, if the difference may be assumed due to sampling fluctuation, the null
hypothesis ( H 0 ) is accepted.
4. Defining the critical rejection regions and making calculations for test statistics
If we select the value of a = Level of significance = 0.05, and use the standard normal
distribution (z-test) as our test statistic for testing the population parameter u, then the
value of the difference between the assumption of null hypothesis (assumed value of the
population parameter) and the value obtained by the analysis of the sample results is not
expected to be more than than 1.96 s at a = 0.05.
Errors in Hypothesis
In statistical hypothesis testing, type I and type II errors are incorrect rejection of
a true null hypothesis or failure to reject a false null hypothesis, respectively. More simply
stated, a type I error is detecting an effect that is not present, while a type II error is
failing to detect an effect that is present. The terms "type I error" and "type II error" are
often used interchangeably with the general notion of false positives and false negatives
in binary classification, such as medical testing, but narrowly speaking refer specifically
to statistical hypothesis testing.
Two types of errors can result from a hypothesis test:
i) Type I error: A Type I error occurs when the researcher rejects a null hypothesis
when it is true. The probability of committing a Type I error is called the
significance level. This probability is also called alpha, and is often denoted by
.
ii) Type II error: A Type II error occurs when the researcher fails to reject a null
hypothesis that is false. The probability of committing a Type II error is called
Beta, and is often denoted by . The probability of not committing a Type II
error is called the Power of the test.
Null Hypothesis
It is used for testing the hypothesis formulated by the researcher. Researchers treat
evidence that supports a hypothesis differently from the evidence that opposes it. They
give negative evidence more importance than to the positive one. It is because the negative
evidence tarnishes the hypothesis. It shows that the predictions made by the hypothesis
are wrong. The null hypothesis simply states that there is no relationship between the
variables or the relationship between the variables is "zero." That is how symbolically null
Notes
hypothesis is denoted as "H0". For example: H 0 = There is no relationship between the
level of job commitment and the level of efficiency. Or H 0 = The relationship between
level of job commitment and the level of efficiency is zero. Or The two variables are
independent of each other. It does not take into consideration the direction of association
(i.e. H 0 is non directional), which may be a second step in testing the hypothesis. First
we look whether or not there is an association then we go for the direction of association
and the strength of association. Experts recommend that we test our hypothesis indirectly
by testing the null hypothesis. In case we have any credibility in our hypothesis then the
research data should reject the null hypothesis. Rejection of the null hypothesis leads
to the acceptance of the alternative hypothesis.
Alternative Hypothesis
The alternative (to the null) hypothesis simply states that there is a relationship
between the variables under study. In our example it could be: there is a relationship
between the level of job commitment and the level of efficiency. Not only there is an
association between the two variables under study but also the relationship is perfect which
is indicated by the number "1". Thereby the alternative hypothesis is symbolically denoted
as " H 1 ". It can be written like this: H 1 : There is a relationship between the level of job
commitment of the officers and their level of efficiency.
4.12 t – test
A statistical examination of two population means. A two-sample t-test examines
whether two samples are different and is commonly used when the variances of two normal
distributions are unknown and when an experiment uses a small sample size.
X
t
Formula: S
N
Where, is the sample mean, Ä is a specified value to be tested, s is the sample
standard deviation and n is the size of the sample. Look up the significance level of the
z-value in the standard normal table.
When the standard deviation of the sample is substituted for the standard deviation
of the population, the statistic does not have a normal distribution; it has what is called
the t-distribution. Because there is a different t-distribution for each sample size, it is not
practical to list a separate area of the curve table for each one. Instead, critical t-values
for common alpha levels (0.10, 0.05, 0.01, and so forth) are usually given in a single table
for a range of sample sizes. For very large samples, the t-distribution approximates the
standard normal (z) distribution. In practice, it is best to use t-distributions any time the
population standard deviation is not known.
Values in the t-table are not actually listed by sample size but by degrees of freedom
(df). The number of degrees of freedom for a problem involving the t-distribution for sample
size n is simply n – 1 for a one-sample mean problem.
Assumptions
Z
Most t-test statistics have the form T , where Z and s are functions of the data.
S
Typically, Z is designed to be sensitive to the alternative hypothesis (i.e. its magnitude
tends to be larger when the alternative hypothesis is true), whereas s is a scaling
parameter that allows the distribution of T to be determined.
X
Z
As an example, in the one-sample t-test , where
X is the sample mean
n
of the data, n is the sample size, and ó is the population standard deviation of the data; S
in the one-sample t-test is / , where is the sample standard deviation.
The assumptions underlying a t-test are that:
i) Z follows a standard normal distribution under the null hypothesis
ii) p 2 follows a 2 distribution with p degrees of freedom under the null hypothesis,
where p is a positive constant
iii) Z and S are independent.
4.13 F – test
F-test is any statistical test in which the test statistic has an F-distribution under
the null hypothesis. It is most often used when comparing statistical models that have
been fitted to a data set, in order to identify the model that best fits the population from
which the data were sampled.
Notes Common examples of the use of F-tests are, for example, the study of the following
cases:
(i) The hypothesis that the means of a given set of normally distributed populations,
all having the same standard deviation, are equal. This is perhaps the best-
known F-test, and plays an important role in the analysis of variance (ANOVA).
(ii) The hypothesis that a proposed regression model fits the data well. See Lack-
of-fit sum of squares.
(iii) The hypothesis that a data set in a regression analysis follows the simpler of
two proposed linear models that are nested within each other.
The F-test is designed to test if two population variances are equal. It does this by
comparing the ratio of two variances. So, if the variances are equal, the ratio of the
variances will be 1.
If the null hypothesis is true, then the F test-statistic given above can be simplified
(dramatically). This ratio of sample variances will be test statistic used. If the null
hypothesis is false, then we will reject the null hypothesis that the ratio was equal to
1 and our assumption that they were equal.
There are several different F-tables. Each one has a different level of significance.
So, find the correct level of significance first, and then look up the numerator degrees
of freedom and the denominator degrees of freedom to find the critical value.
You will notice that all of the tables only give level of significance for right tail tests.
Because the F distribution is not symmetric and there are no negative values, you may
not simply take the opposite of the right critical value to find the left critical value. The
way to find a left critical value is to reverse the degrees of freedom, look up the right critical
value, and then take the reciprocal of this value. For example, the critical value with 0.05
on the left with 12 numerator and 15 denominator degrees of freedom is found of taking
the reciprocal of the critical value with 0.05 on the right with 15 numerator and 12
denominator degrees of freedom.
F Test Formula
F Test is a method to compare variance of two different set of values. F test is applied
on F distribution under null hypothesis. For calculating F test value, we first find the mean
of two given observations and then calculate their variance. F test value is expressed as
the ratio of variances of two observations. The comparison between the variances of two
sets of data can lead to many predictions. The formula for F test is given below:
variance 1 12
F value
variance 2 22
Variance is given by the following formula:
( x x )2
2
n 1
Where,
2 = Variance
x = Values given in a set of data
The chi-squared test, when used with the standard approximation that a chi-squared
distribution is applicable, has the following assumptions:
i) Simple random sample: The sample data is a random sampling from a fixed
distribution or population where each member of the population has an equal
probability of selection. Variants of the test have been developed for complex
samples, such as where the data is weighted.
ii) Sample size (whole table): A sample with a sufficiently large size is assumed.
If a chi squared test is conducted on a sample with a smaller size, then the
chi squared test will yield an inaccurate inference. The researcher, by using
chi squared test on small samples, might end up committing a Type II error.
iii) Expected cell count: Adequate expected cell counts. Some require 5 or more,
and others require 10 or more. A common rule is 5 or more in all cells of a
2-by-2 table, and 5 or more in 80% of cells in larger tables, but no cells with
zero expected count. When this assumption is not met, Yates’s correction is
applied.
iv) Independence: The observations are always assumed to be independent of
each other. This means chi-squared cannot be used to test correlated data (like
matched pairs or panel data). In those cases you might want to turn to
McNamara’s test.
F bO g IJ
2
G
Ei
GH JK
2 i
x
Ei
4.15 Summary
Probability is the likelihood that an event will occur and is calculated by dividing the
number of favourable outcomes by the total number of possible outcomes.
Experiment refers to describe an act which can be repeated under some given
conditions. Random experiments are those experiments whose results depend on chance
such as tossing a coin, throwing a dice etc. The outcomes of random experiment are
events.
A probability distribution is a statistical function. It describes all the possible values
and possibilities that a random variable can take within a given range. This range will be
in between the minimum and maximum statistically possible values. But, it is where the
possible value is likely to be plotted on the probability distribution depends on a number
of factors. These factors include the distribution's mean, standard deviation, skewness
and kurtosis.
The Poisson distribution is used to describe discrete quantitative data such as counts
in which the populations size n is large, the probability of an individual event is small,
but the expected number of events, n is moderate say five or more.
Hypergeometric distribution is a probability distribution that’s very similar to the
binomial distribution. In fact, the binomial distribution is a very good approximation of the
hypergeometric distribution as long as you are sampling 5% or less of the population.
Non-probability samples that are unrestricted are called convenient sampling.
Convenience sampling refers to the collection of information from members of population
who are conveniently available to provide it. Researchers or field workers have the freedom
to choose as samples whomever they find, thus it is named as convenience.
Judgment sampling is a non-probability sampling method in which participants are
selected according to an experienced individual’s belief that they will meet the
requirements of the study. The researcher selects sample members who conform to some Notes
criterion.
Snowball sampling is a non-probability sampling method in which a set of
respondents are chosen who help the researcher to identify additional respondents to be
included in the study. This method of sampling is also called as referral sampling because
one respondent refers other potential respondents. This method involves probability and
non-probability methods.
Hypothesis test is a method of making decisions using data from a scientific study.
In statistics, a result is called statistically significant if it has been predicted as unlikely
to have occurred by chance alone, according to a pre-determined threshold probability,
the significance level.
F-test is any statistical test in which the test statistic has an F-distribution under
the null hypothesis. It is most often used when comparing statistical models that have
been fitted to a data set, in order to identify the model that best fits the population from
which the data were sampled.
Chi-square test is a test that uses the chi-square statistic to test the fit between
a theoretical frequency distribution and a frequency distribution of observed data for which
each observation may fall into one of several classes.
4.22 Bibliography
1. Srivastava V. K. etal – Quantitative Techniques for Managerial Decision Making,
Wiley Eastern Ltd
2. Richard, I.Levin and Charles A.Kirkpatrick – Quantitative Approaches to
Management, McGraw Hill, Kogakusha Ltd.
3. Prem.S.Mann, Introductory Statistics, 7th Edition, Wiley India, 2016.
Notes 4. Aczel A.D. and Sounderpandian J., “Complete Business Statistics”, 6th edition,
Tata McGraw Hill Publishing Company Ltd., New Delhi, 2012.
5. Anderson D.R., Sweeney D.J. and Williams T.A., Statistics for business and
economics, 11th edition, Thomson (South – Western) Asia, Singapore, 2012.
6. Budnik, Frank S Dennis Mcleaavey, Richard Mojena – Principles of Operation
Research - AIT BS New Delhi.
7. Sharma J K – Operation Research- theory and applications-Mc Millan,New Delhi
8. Kalavathy S. – Operation Research – Vikas Pub Co
9. Gould F J – Introduction to Management Science – Englewood Cliffs N J Prentice
Hall.
10. Naray J K, Operation Research, theory and applications – Mc Millan, New Dehi.
11. Taha Hamdy, Operations Research, Prentice Hall of India
12. Tulasian: Quantitative Techniques: Pearson Ed.
13. Vohr.N.D. Quantitative Techniques in Management, TMH.
11. Stevenson W.D, Introduction to Management Science, TMH.