0% found this document useful (0 votes)
2 views

ML NOTES

Factor analysis is a statistical method used to identify underlying structures in complex datasets by condensing observed variables into unobserved factors. It includes exploratory and confirmatory types, with applications in social sciences, market research, business, healthcare, and education. Additionally, dimension reduction techniques enhance data analysis by simplifying datasets while retaining essential information.

Uploaded by

laxmi
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
2 views

ML NOTES

Factor analysis is a statistical method used to identify underlying structures in complex datasets by condensing observed variables into unobserved factors. It includes exploratory and confirmatory types, with applications in social sciences, market research, business, healthcare, and education. Additionally, dimension reduction techniques enhance data analysis by simplifying datasets while retaining essential information.

Uploaded by

laxmi
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 12

What is Factor Analysis?

Factor analysis is a potent statistical method for comprehending complex datasets’ underlying
structure or patterns. Its primary objective is to condense many observed variables into a smaller set
of unobserved variables called factors. These factors aim to capture the essential information from
the original variables, simplifying the understanding and interpretation of data

Key Concepts of Factor Analysis

Factor analysis hinges on several pivotal concepts underpinning its functionality and application
across diverse domains. Understanding these core concepts is fundamental to grasping the essence
of this statistical technique.

1. Variables, Factors, and Observed Data

• Variables: These are the measurable quantities or items used in an analysis, such as survey
responses, test scores, or economic indicators.

• Factors: Unobservable latent variables representing underlying constructs or dimensions


influencing the observed variables’ behaviour.

• Observed Data: The data matrix containing measurements or responses across multiple
variables for each observation or individual.

2. Types of Factor Analysis

• Exploratory factor analysis (EFA) explores underlying relationships between observed


variables and identifies potential factors without preconceived notions about the structure.

• Confirmatory Factor Analysis (CFA): Validates pre-existing theories or hypotheses about the
structure of relationships among variables by testing and confirming a specific factor
structure.

Factor Analysis (FA)

• Objective: FA aims to identify latent factors that underlie the observed variables. It assumes
that these unobserved factors influence observed variables and are related to each other
and the observed variables.

• Methodology: FA seeks to explain the covariance between observed variables in terms of


underlying latent factors. It’s more concerned with uncovering the structure and
relationships between variables.

• Assumptions: FA assumes that a smaller number of latent factors influences observed


variables and that measurement errors are present in the observed variables.

• Usage: It is often used in social sciences, psychology, and market research to identify
underlying constructs, understand relationships between variables, and uncover hidden
patterns in data.

Applications of Factor Analysis

Factor Analysis finds wide-ranging applications across numerous fields due to its ability to unveil
underlying structures within complex datasets. Some prominent areas where Factor Analysis is
extensively utilized include:
See also K-Nearest Neighbours Explained, Practical Guide & How To Tutorial In Python

1. Social Sciences

• Psychology: Identifying latent constructs such as personality traits, attitudes, or cognitive


abilities from survey responses or behavioural data.

• Sociology: Studies social phenomena by analyzing survey patterns related to societal


attitudes, behaviours, or cultural factors.

2. Market Research and Consumer Behavior

• Segmentation: Identifying consumer segments based on purchasing patterns, preferences,


or demographic variables.

• Product Development: Determining underlying attributes influencing consumer perceptions


or preferences for product design and marketing strategies.

3. Business and Economics

• Financial Analysis: Reducing many financial indicators into key underlying factors influencing
market performance or economic trends.

• Human Resources: Identifying underlying factors impacting employee satisfaction,


engagement, or performance.

4. Healthcare and Medicine

• Clinical Research: Uncovering latent variables related to symptoms, disease progression, or


treatment effectiveness.

• Health Behavior Studies: Analyzing patterns in health-related behaviors, risk factors, or


patient-reported outcomes.

5. Education and Testing

• Educational Assessment: Understanding factors affecting academic performance, learning


styles, or educational outcomes.

• Test Development: Validating test items and determining underlying constructs measured by
assessments.

6. Environmental and Social Sciences

• Environmental Studies: Identifying factors contributing to environmental attitudes,


behaviours, or sustainability initiatives.

• Opinion Polls and Surveys: Analyzing public opinions or perceptions on social, political, or
environmental issues.
Dimension Reduction-

In pattern recognition, Dimension Reduction is defined as-

• It is a process of converting a data set having vast dimensions into a data set with lesser
dimensions.

• It ensures that the converted data set conveys similar information concisely.

Example-

Consider the following example-

• The following graph shows two dimensions x1 and x2.

• x1 represents the measurement of several objects in cm.

• x2 represents the measurement of several objects in inches.

In machine learning,

• Using both these dimensions convey similar information.

• Also, they introduce a lot of noise in the system.

• So, it is better to use just one dimension.

Using dimension reduction techniques-

• We convert the dimensions of data from 2 dimensions (x1 and x2) to 1 dimension (z1).
• It makes the data relatively easier to explain.

Benefits-

Dimension reduction offers several benefits such as-

• It compresses the data and thus reduces the storage space requirements.

• It reduces the time required for computation since less dimensions require less computation.

• It eliminates the redundant features.

• It improves the model performance.

Least Square Method | Definition Graph and Formula

Last Updated : 20 Aug, 2024

Least Square method is a fundamental mathematical technique widely used in data analysis,
statistics, and regression modeling to identify the best-fitting curve or line for a given set of data
points. This method ensures that the overall error is reduced, providing a highly accurate model for
predicting future data trends.

In statistics, when the data can be represented on a cartesian plane by using the independent and
dependent variable as the x and y coordinates, it is called scatter data. This data might not be useful
in making interpretations or predicting the values of the dependent variable for the independent
variable. So, we try to get an equation of a line that fits best to the given data points with the help
of the Least Square Method.

In this article, we will learn the least square method, its formula, graph, and solved examples on it.

Table of Content

• What is the Least Square Method?

• Formula for Least Square Method

• Least Square Method Graph

o Limitations of the Least Square Method

• Least Square Method Solved Examples


• How Do You Calculate Least Squares?

What is the Least Square Method?

Least Square Method is used to derive a generalized linear equation between two variables. when
the value of the dependent and independent variable is represented as the x and y coordinates in a
2D cartesian coordinate system. Initially, known values are marked on a plot. The plot obtained at
this point is called a scatter plot.

Then, we try to represent all the marked points as a straight line or a linear equation. The equation
of such a line is obtained with the help of the Least Square method. This is done to get the value of
the dependent variable for an independent variable for which the value was initially unknown. This
helps us to make predictions for the value of dependent variable.

Least Square Method Definition

Least Squares method is a statistical technique used to find the equation of best-fitting curve or line
to a set of data points by minimizing the sum of the squared differences between the observed values
and the values predicted by the model.

This method aims at minimizing the sum of squares of deviations as much as possible. The line
obtained from such a method is called a regression line or line of best fit.
Formula for Least Square Method

Least Square Method formula is used to find the best-fitting line through a set of data points. For a
simple linear regression, which is a line of the form y=mx+c, where y is the dependent variable, x is
the independent variable, a is the slope of the line, and b is the y-intercept, the formulas to calculate
the slope (m) and intercept (c) of the line are derived from the following equations:

1. Slope (m) Formula: m = n(∑xy)−(∑x)(∑y) / n(∑x2)−(∑x)2

2. Intercept (c) Formula: c = (∑y)−a(∑x) / n


Where:

• n is the number of data points,

• ∑xy is the sum of the product of each pair of x and y values,

• ∑x is the sum of all x values,

• ∑y is the sum of all y values,

• ∑x2 is the sum of the squares of x values.

The steps to find the line of best fit by using the least square method is discussed below:

• Step 1: Denote the independent variable values as xi and the dependent ones as yi.

• Step 2: Calculate the average values of xi and yi as X and Y.

• Step 3: Presume the equation of the line of best fit as y = mx + c, where m is the slope of the
line and c represents the intercept of the line on the Y-axis.

• Step 4: The slope m can be calculated from the following formula:

m = [Σ (X – xi)×(Y – yi)] / Σ(X – xi)2

• Step 5: The intercept c is calculated from the following formula:

c = Y – mX

Thus, we obtain the line of best fit as y = mx + c, where values of m and c can be calculated from the
formulae defined above.

These formulas are used to calculate the parameters of the line that best fits the data according to
the criterion of the least squares, minimizing the sum of the squared differences between the
observed values and the values predicted by the linear model.

Least Square Method Graph

Let us have a look at how the data points and the line of best fit obtained from the Least Square
method look when plotted on a graph.
The red points in the above plot represent the data points for the sample data
available. Independent variables are plotted as x-coordinates and dependent ones are plotted as y-
coordinates. The equation of the line of best fit obtained from the Least Square method is plotted as
the red line in the graph.

We can conclude from the above graph that how the Least Square method helps us to find a line
that best fits the given data points and hence can be used to make further predictions about the
value of the dependent variable where it is not known initially.

Limitations of the Least Square Method

The Least Square method assumes that the data is evenly distributed and doesn’t contain any
outliers for deriving a line of best fit. But, this method doesn’t provide accurate results for unevenly
distributed data or for data containing outliers.
E
MARKOV CHAIN MONTE CARLO METHODS

REINFORCEMENT LEARNING

You might also like