TS&PDA
TS&PDA
Answer:
Binary choice models are statistical models used to analyze situations where the outcome
variable can take on only two possible values, typically coded as 0 and 1. These models are
widely used in various fields, including economics, finance, marketing, and social sciences.
ii. Understanding Factors Influencing Choices: Binary choice models help identify the
factors that influence individuals' decisions. By analyzing the relationship between
these factors and the binary outcome, researchers can gain insights into the underlying
decision-making process.
iii. Making Predictions: These models can be used to predict the probability of an
individual making a particular choice based on their characteristics and the factors
influencing their decision.
iv. Policy Evaluation: Binary choice models can be used to evaluate the effectiveness of
policies or interventions aimed at influencing individual choices.
1. Logit Model: This is one of the most commonly used binary choice models. It assumes
that the probability of the outcome being 1 follows a logistic distribution.
2. Probit Model: Similar to the logit model, but it assumes that the probability of the
outcome being 1 follows a standard normal distribution.
3. Linear Probability Model (LPM): While simpler, the LPM has limitations as it can
predict probabilities outside the range of 0 to 1.
2. What is Linear Probability Model (LPM)? What are the limitations of LPM?
Answer:
The Linear Probability Model (LPM) is a statistical method used to model the relationship
between a binary dependent variable (one that can only take two values, typically 0 or 1) and
one or more independent variables. It's a simple approach where the probability of the
dependent variable being 1 is modeled as a linear function of the independent variables.
Limitations of LPM
1. Predicted Probabilities Outside the Range of 0-1: The biggest issue with LPM is that it
can produce predicted probabilities that are less than 0 or greater than 1. Since
probabilities must lie between 0 and 1, this makes the model's predictions unrealistic
and difficult to interpret.
2. Heteroscedasticity: The variance of the error term in LPM is not constant, violating the
assumption of homoscedasticity required for valid statistical inference. This can lead to
unreliable estimates of the model's parameters and standard errors.
3. Non-Normality of Errors: The errors in LPM are not normally distributed, another
assumption violated by the model. This can affect the accuracy of hypothesis tests and
confidence intervals.
4. Questionable R2
Answer:
Odds and log of odds are related concepts used to express the likelihood of an event.
Odds are a way of expressing the likelihood of an event occurring. They represent the ratio of
the probability of the event happening to the probability of the event not happening.
Log of Odds: The log of odds is the natural logarithm of the odds. It's a transformation used
in statistical modeling, particularly in logistic regression.3
o Linearity: In logistic regression, the log of odds has a linear relationship with the
predictor variables, making it easier to model.
o Range: The log of odds can take any value, unlike probability which is restricted
between 0 and 1.
4. What is a Logit model? Explain some of the features of the logit model?
Answer:
The Logit Model is a statistical model used to analyze the relationship between a binary
dependent variable (one that can take on only two values, typically 0 or 1) and one or more
independent variables.1 It's a popular alternative to the Linear Probability Model (LPM) and
overcomes many of its limitations.
i. Based on Logistic Distribution: The Logit Model assumes that the probability of the
dependent variable being 1 follows a logistic distribution. This ensures that the
predicted probabilities always fall within the range of 0 to 1, unlike the LPM.
ii. Uses Log-Odds: The Logit Model models the log-odds of the event occurring.3 This
transformation allows for a linear relationship between the log-odds and the
independent variables, making it easier to estimate and interpret the model.
iii. Estimates Probabilities: The Logit Model estimates the probability of the dependent
variable being 1 for different values of the independent variables. This allows for
predictions and insights into the likelihood of the event occurring.
iv. Handles Heteroscedasticity: The Logit Model inherently addresses the issue of
heteroscedasticity, which is a common problem in the LPM.
v. Widely Applicable: The Logit Model has a wide range of applications in various fields,
including economics, finance, marketing, social sciences, and medicine.
5. Why Logit and Probit model differs from the LPM model?
Answer:
The Logit and Probit models offer significant advantages over the Linear Probability Model
(LPM) when dealing with binary dependent variables (those that can only take on two values,
like 0 or 1). Here's a breakdown of their key differences:
1. Predicted Probabilities:
• LPM: Can predict probabilities outside the 0-1 range. This is a major limitation as
probabilities must lie between 0 and 1.
• Logit & Probit: Ensure that predicted probabilities always fall within the 0-1 range.
This is because they are based on underlying distributions (logistic for Logit, normal
for Probit) that are bounded within these limits.
2. Heteroscedasticity:
• LPM: Suffers from heteroscedasticity, meaning the variance of the error term is not
constant across different values of the independent variables.1 This violates a key
assumption of linear regression.
• Logit & Probit: Inherently address the issue of heteroscedasticity, providing more
robust and reliable estimates.
3. Underlying Distributions:
• LPM: Assumes a linear relationship between the independent variables and the
probability of the outcome, which may not always hold in reality.
• Logit: Assumes that the probability of the outcome follows a logistic distribution.
• Probit: Assumes that the probability of the outcome follows a standard normal
distribution.
4. Interpretation of Coefficients:
• LPM: Coefficients directly represent the change in the probability of the outcome for
a one-unit change in the independent variable. However, as mentioned earlier, this can
lead to unrealistic predictions.
• Logit & Probit: Coefficients represent the change in the log-odds of the outcome for
a one-unit change in the independent variable.2 These coefficients are not as easily
interpretable as in LPM, but they provide a more accurate representation of the
relationship between the variables.
6. What is the probit model? How is it different from the logit model? Show
graphically & theoretically?
Answer:
The Probit Model is a statistical method used to analyze the relationship between a binary
dependent variable (one that can take on only two values, typically 0 or 1) and one or more
independent variables. It's a popular alternative to the Logit model and the Linear Probability
Model (LPM).
Differences between Probit and Logit Models
1. Underlying Distribution:
2. Link Function:
• Probit: Uses the probit link function (inverse of the standard normal CDF).
• Logit: The S-curve of the Logit model has slightly heavier tails than the Probit model.
This means it can handle extreme values in the predictors or outcomes slightly better.
• Probit: The S-curve of the Probit model is steeper near the center and approaches the
axes more quickly than the Logit model. This implies that the Probit model might be
less accommodating of outliers compared to the Logit model.
Applied Part
Answer:
a) Factors influencing Homeownership:
Several factors can significantly impact whether a household owns a house or not. Here are
some key ones:
• Household Size & Composition: Larger families or those with children may have a
greater need for space and thus be more likely to own a house.
• Education Level: Higher education levels are often associated with higher incomes
and, consequently, a greater likelihood of homeownership.
• Age of Household Head: Older individuals are more likely to have accumulated
wealth and, therefore, own a house.
where:
Interpretation:
• The intercept (-0.9457) suggests that for a household with zero income, the predicted
probability of homeownership is negative. This is not realistic in the context of
homeownership and highlights a limitation of the linear probability model (LPM).
• The coefficient of family income (0.1021) indicates that for every $1,000 increase in
income, the predicted probability of homeownership is expected to increase by
0.1021.
Statistical Significance:
Both coefficients have t-statistics that are much larger than the critical values for a typical
significance level (e.g., 0.05). This suggests that both the intercept and the slope coefficient
are statistically significant, meaning they are unlikely to have occurred by chance.
d) Interpretation if X=15:
This means that a household with an income of $15,000 has a predicted probability of
homeownership of approximately 64.68%.
In the LPM, the regressand (predicted probability) is not guaranteed to be within the 0-to-1
range. This is a major drawback of the LPM. In cases where the predicted probability falls
outside this range, the model's predictions become unrealistic.
f) Heteroscedasticity:
Heteroscedasticity refers to the situation where the variance of the error term is not constant
across1 different values of the explanatory variable(s). In the LPM, heteroscedasticity is often
a concern because the variance of the error term tends to be larger for extreme values of the
predicted probabilities (close to 0 or 1).
One common method to address heteroscedasticity in the LPM is by using Weighted Least
Squares (WLS). WLS assigns different weights to observations based on their predicted
probabilities, giving more weight to observations with lower predicted probabilities and less
weight to those with higher predicted probabilities.
The marginal effect in the LPM is the coefficient of the explanatory variable (income in this
case). It represents the change in the predicted probability of homeownership for a one-unit
change in the explanatory variable.
In this model, the marginal effect of income is 0.1021, indicating that a $1,000 increase in
income is associated with a 0.1021 increase in the predicted probability of homeownership.
Answer:
• Both the intercept and the income coefficient have very high t-statistics (in absolute
value) and are statistically significant at conventional levels (e.g., 5% or 1%). This
indicates that the estimated coefficients are unlikely to be zero by chance.
• Income Coefficient: The coefficient of 0.0792 indicates that for every $1,000
increase in family income, the estimated log-odds of homeownership are expected to
increase by 0.0792.
▪ exp(0.0792) ≈ 1.082
o This means that for every $1,000 increase in income, the odds of
homeownership are estimated to increase by a factor of 1.082 (or by about
8.2%).
The marginal effect in a Logit model is not constant. It depends on the value of the
independent variable (X).
• Heteroscedasticity: LPM assumes constant variance of the error term, which is often
violated in practice.
• Non-normality of Errors: The errors in LPM are not normally distributed, which can
affect the validity of statistical inferences.
Answer:
a. Interpretation of Coefficients:
• Intercept (1.1387): This represents the log-odds of a high murder rate when all the
independent variables (population size, growth rate, and reading quotient) are zero.
However, this interpretation might not be practically meaningful in this context as it's
unlikely to have an SMSA with zero population or growth rate.
• Population Growth Rate (0.0561): This coefficient indicates that a one-unit increase
in the population growth rate is associated with a 0.0561 increase in the log-odds of a
high murder rate, holding other variables constant.
To determine statistical significance, we can use the t-statistic, which is calculated as the
coefficient divided by its standard error. As a rule of thumb, if the absolute value of the t-
statistic is greater than 2, we can generally consider the coefficient statistically significant at
the 5% level.
Therefore, the coefficients for population growth rate and reading quotient are statistically
significant at the 5% level.
• exp(-0.4050) ≈ 0.667
This means that a one-unit increase in the reading quotient is associated with a decrease in the
odds of a high murder rate by a factor of approximately 0.667.
Since the population growth rate is already in percentage points, the coefficient itself
represents the effect on the log-odds for a one percentage point increase.
• exp(0.0561) ≈ 1.057
This means that a one percentage point increase in the population growth rate is associated
with an increase in the odds of a high murder rate by a factor of approximately 1.057.
Answer:
Bernoulli Distribution: A Bernoulli random variable, often denoted as X, can only take on
two values: 0 or 1. It represents a single trial with two possible outcomes: success (X=1) with
probability p, and failure (X=0) with probability (1-p).
o E[X] = (0 * (1-p)) + (1 * p) = p
Variance (Var(X)):
o Var(X) = p - p^2
o Var(X) = p(1-p)
This makes sense intuitively. The variance is highest when p = 0.5 (meaning success and
failure are equally likely), and it decreases as p gets closer to 0 or 1.
Answer:
P(Y = 1) = 1 / (1 + exp(-Z))
where:
• β₀ is the intercept
• β₁, β₂, ..., βₖ are the coefficients for the independent variables X₁, X₂, ..., Xₖ
Marginal Effect
The marginal effect of a variable (let's say X₁) represents the change in the probability of the
event (Y = 1) for a one-unit increase in that variable, holding other variables constant.
Derivation
o Let P(Y = 1) = P
o P = 1 / (1 + exp(-Z))
o Rearranging:
▪ 1/P - 1 = exp(-Z)
▪ exp(Z) = (1/P) - 1
▪ Z = ln((1/P) - 1)
o dZ/dX₁ = β₁
3. Differentiating P with respect to Z:
o dP/dZ = P * (1 - P)
o This can be derived using the chain rule and the properties of the logistic
function.
o dP/dX₁ = P * (1 - P) * β₁
Marginal Effect = β₁ * P * (1 - P)