0% found this document useful (0 votes)
4 views

Statistical Methods

Uploaded by

indubaijaiwal
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
4 views

Statistical Methods

Uploaded by

indubaijaiwal
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 15

Averages

Definition: Averages are statistical measures that summarize or represent a group of


data with a single value. They are used to identify a central or typical value in a dataset.

Requisites of a Good Average:


1. Simplicity: It should be easy to understand and calculate.
2. Representativeness: The average should adequately represent the entire
dataset.
3. Rigidity: The value of the average should not change with minor variations in the
data.
4. Based on All Observations: It should consider all the data points to provide a
comprehensive summary.
5. Susceptibility to Sampling: It should be applicable to all types of data and
consistent across different samples.
6. Mathematical Properties: It should lend itself to further mathematical treatment.
7. Least Affected by Extreme Values: Ideally, it should not be overly influenced by
outliners.

Types of Averages:

1. Mean (Arithmetic Mean):

Definition: The sum of all observations divided by the total number of observations.

Formula:

Advantages: Simple to compute, uses all data points.


Disadvantages: Sensitive to outliers.
2. Median:

Definition: The middle value when data is arranged in ascending or descending order.
If there is an even number of observations, it is the average of the two middle values.

Formula:

Advantages: Not affected by extreme values.

Disadvantages: Does not use all data points.

3. Mode:

Definition: The value that occurs most frequently in the dataset.

Advantages: Useful for categorical data.


Disadvantages: May not be unique or may not exist in some datasets.

Dispersion
Definition: Dispersion refers to the extent to which data values are spread out or
scattered. It measures the variation or diversity within a dataset.
Significance of Measuring Variation:

1. Understanding Data Spread: Helps in assessing the consistency or reliability of


the data.
2. Comparative Analysis: Allows comparison of variability between different
datasets.
3. Decision Making: Facilitates better decision-making by providing insights into
risk and uncertainty.
4. Identifying Outliers: Highlights extreme values that may influence results.
5. Basis for Further Analysis: Essential for advanced statistical analysis, such as
hypothesis testing.

Methods of Studying Variation:

1. Mean Deviation:

Definition: The average of the absolute deviations of each data point from the mean or
median.

Formula:

Advantages: Easy to compute and interpret.

Disadvantages: Ignores the direction of deviations (positive or negative).

2. Standard Deviation:

Definition: The square root of the average of the squared deviations from the mean. It
provides a precise measure of dispersion.

Formula:
Advantages: Widely used, incorporates all data points, and has mathematical
significance.

Disadvantages: Sensitive to extreme values.

Correlation Analysis

Significance of Studying Correlation


Correlation analysis helps in determining the strength and direction of the relationship
between two variables. Its significance includes:

1. Prediction: Helps in predicting the value of one variable based on the other.

2. Decision Making: Useful in business and economics for understanding relationships


like demand and supply, income and expenditure, etc.

3. Identifying Dependencies: Helps identify dependent and independent variables.

4. Understanding Relationships: Provides insights into the nature of relationships,


aiding researchers and policymakers.

Types of Correlation :
1. Positive Correlation: Both variables increase or decrease together (e.g., income
and expenditure).

2. Negative Correlation: One variable increases while the other decreases (e.g., price
and demand).

3. Zero Correlation: No relationship exists between the variables.


4. Linear Correlation: Variables change proportionally.
5. Non-linear Correlation: Variables do not change proportionally but still exhibit a
pattern.

Methods of Studying Correlation:


1. Karl Pearson’s Coefficient of Correlation (r):

A quantitative measure of the strength and direction of the linear relationship between
two variables.

Formula:

2. Rank Correlation Coefficient (Spearman’s ):


Measures the relationship between two ranked variables.

Formula:
Use of Regression Analysis:
1. Predictive Modeling: Used to predict the value of one variable based on another.

2. Understanding Relationships: Analyzes how a dependent variable is influenced by


one or more independent variables.

3. Optimizing Decision-Making: Helps businesses in forecasting, inventory control,


and risk analysis.

Difference Between Correlation and Regression Analysis:

Two Regression Equations:


1. Regression Equation of Y on X:

Y = a + bX

2. Regression Equation of X on Y:

X = a' + b'Y

These equations describe how the dependent variable (Y) changes with changes in the
independent variable (X) and vice versa.
Time Series Analysis

1. Utility of Time Series Analysis:


Time Series Analysis involves studying historical data points collected or recorded at
specific intervals. Its utilities include:

Trend Analysis: Identifying long-term trends in data for strategic planning.

Forecasting: Making future predictions based on historical patterns.

Understanding Variability: Analyzing seasonal and cyclical variations to optimize


operations.

Decision-Making: Supporting informed decisions in areas like sales, production, and


inventory.
Performance Evaluation: Monitoring performance over time for growth or decline.

2. Components of Time Series:


A time series comprises the following components:

Trend (T): The long-term direction of the data (upward, downward, or constant).

Seasonality (S): Regular, periodic fluctuations (e.g., monthly or quarterly patterns).

Cyclic Variations (C): Irregular fluctuations over time due to economic or business
cycles, often longer than a year.

Irregular or Random Variations (I): Unpredictable, short-term variations caused by


events like natural disasters or strikes.

3. Methods of Measurement:
a) Moving Averages:

This method smoothens time series data to identify trends by averaging data over a
specific period.

Simple Moving Average (SMA): Calculates the unweighted mean of the last ‘n’
observations.
Weighted Moving Average (WMA): Assigns more weight to recent observations,
reflecting their significance.

Steps to Calculate SMA:


1. Select the period (e.g., 3, 5, or 7 observations).
2. Add the values of observations within that period.
3. Divide the total by the number of observations in the period.
4. Shift the period forward and repeat.

b) Method of Least Squares:

This statistical method fits a straight line or curve to time series data to represent the
trend mathematically. The line minimizes the sum of squared deviations of observed
values from the predicted values.
Business Forecasting

1. Steps in Forecasting:
a. Define Objectives: Identify the purpose and scope of forecasting.
b. Analyze Past Data: Collect and study historical data for patterns and trends.
c. Identify Influencing Factors: Consider internal and external factors (economic,
political, etc.).
d. Select Forecasting Method: Choose an appropriate method based on data
characteristics.
e. Develop the Forecast: Apply the chosen method to generate predictions.
f. Validate the Forecast: Compare predictions with actual results to assess
accuracy.
g. Monitor and Update: Continuously refine forecasts based on new data and
outcomes.

2. Methods of Forecasting:
Forecasting methods can be classified into:

a) Qualitative Methods: Based on expert opinions and judgment.


b) Delphi Method: Structured expert consensus.
c) Market Research: Surveys and interviews.
d) Scenario Building: Developing potential future scenarios.
e) Quantitative Methods: Based on mathematical and statistical techniques.
f) Time Series Analysis: Analyzing historical data to predict future values.
g) Regression Analysis: Establishing relationships between variables for
prediction.
h) Exponential Smoothing: Weighting recent observations more heavily for
prediction.

3. Theories of Business Forecasting:


A. Economic Theory: Assumes business activity aligns with economic cycles (e.g.,
GDP growth).
B. Market Theory: Predicts based on supply-demand equilibrium.
C. Technological Theory: Assumes innovations and technology drive business
changes.
D. Managerial Judgment: Relies on managerial experience and intuition.

Probability
1. Three Approaches to Probability:
o Classical Probability: Based on the assumption of equally likely
outcomes.

 Empirical Probability: Based on experimentation or historical data.

 Axiomatic Probability: Based on axioms introduced by Kolmogorov, defining


probability as a measure satisfying certain conditions
 (e.g., ).

2. Addition Theorem of Probability:

3. Multiplication Theorem of Probability:

4. Bayes' Theorem:
5. Probability Distributions:

Intervention in Markets

1. Price Controls:
o Price Ceiling: Maximum price set by the government, typically below
equilibrium (e.g., rent control).
o Price Floor: Minimum price set by the government, typically above
equilibrium (e.g., minimum wage).
2. Support Price:
o A form of price floor where the government guarantees a minimum price
for goods (e.g., agricultural products).
3. Prevention and Control of Monopolies:
o Government policies to regulate or break up monopolies to ensure fair
competition (e.g., antitrust laws, merger controls).
4. System of Dual Pricing:
o Goods are sold at two prices: subsidized prices for essential needs and
higher market prices for others. Example: Rationing systems in developing
countries.
Statistical Inference
Procedure of Testing a Hypothesis:

Hypothesis testing is a formal process to determine whether to accept or reject a


hypothesis based on sample data. The steps are as follows:

1. Formulate the Null (H0) and Alternative Hypotheses (Ha)


o Null Hypothesis (H0): Assumes no effect or no difference.
o Alternative Hypothesis (Ha): Contradicts H0 and states an effect or
difference.

2. Choose a Significance Level (α\alphaα):


o Commonly used levels are 0.05 (5%) or 0.01 (1%).
o α is the probability of rejecting H0 when it is true.

3. Select a Test Statistic:


o Based on the type of data and distribution (e.g., t-test, F-test, χ2-test).

4. Define the Decision Rule:


o Determine the critical value(s) or p-value for the test.

5. Collect Data and Perform the Test:


o Calculate the test statistic using sample data.

6. Make a Decision:
o Compare the test statistic to the critical value or the p-value to α.
o Reject H0 if the test statistic falls in the rejection region or p<α.

7. Interpret the Results:


o State the conclusion in the context of the problem.
Two Types of Errors in Hypothesis Testing:

1. Type I Error (α):


o Rejecting a true null hypothesis.
o Example: Convicting an innocent person.
o Probability of Type I Error = Significance Level (α).

2. Type II Error (β):


o Failing to reject a false null hypothesis.
o Example: Letting a guilty person go free.
o Probability of Type II Error = Depends on the power of the test (1−β).

Two-Tailed and One-Tailed Tests of Hypothesis

1. Two-Tailed Test:

 Tests for any significant difference (either higher or lower).

 Rejection region is on both ends of the distribution.

2. One-Tailed Test:
 Tests for a specific direction (greater or smaller).
Tests in Statistical Inference:
t-Test :

 Purpose: Compares the means of two groups (small sample sizes).


 Types:
1. One-Sample t-Test: Compares the sample mean to a known value.

2. Two-Sample t-Test: Compares means of two independent groups.


3. Paired t-Test: Compares means of related groups (e.g., before and after).
 Assumptions:
o Data is approximately normally distributed.
o Variances are equal (for two-sample t-tests).

F-Test :

 Purpose: Compares the variances of two groups.


 Formula:

 Applications:
o Testing the equality of variances.
o Used in Analysis of Variance (ANOVA).
 Assumptions:
o Samples are independent.
o Data follows a normal distribution.
Chi-Square (χ2) Test :

 Purpose: Tests for relationships between categorical variables.


 Types:
1. Test of Independence: Determines if two categorical variables are
independent.
2. Test of Goodness-of-Fit: Determines if data fits a specific distribution.
 Formula

 Assumptions:
o Observations are independent.
o Expected frequencies are at least 5.

Analysis of Variance (ANOVA) :

 Purpose: Compares means of three or more groups.


 Types:
1. One-Way ANOVA: Tests the effect of one independent variable.
2. Two-Way ANOVA: Tests the effect of two independent variables.
 Key Concepts:
o Total Variation = Variation Between Groups + Variation Within Groups.
 Test Statistic

 Assumptions:
o Data is normally distributed.
o Variances are equal.
o Observations are independent.

You might also like