PA
PA
In order to allow for non-linear effects a GAM replaces each linear component
βjxj with a smooth non-linear function fj(xj)
This is called an additive model because we estimate each fj(xj) for j=1,2,3,....,p
and then add together all of these individual contributions.
1. Generalized Additive Models (GAMs):**Description: GAMs extend
generalized linear models by allowing non-linear relationships between
predictors and the response variable using smooth functions.**Form:
g(E(Y))=β0+f1(X1)+f2(X2)+⋯+fn(Xn)g(E(Y)) = \beta_0 + f_1(X_1) + f_2(X_2) + \
cdots + f_n(X_n)g(E(Y))=β0+f1(X1)+f2(X2)+⋯+fn(Xn) where ggg is a link
function, β0\beta_0β0 is the intercept, and fif_ifi are smooth functions of the
predictors.**Use Case: Useful when the relationship between predictors and
the response is not strictly linear but can be modeled with smooth functions.
2. Additive Models (AMs):**Description: AMs are a subset of GAMs where the
predictors are added linearly, but each predictor can be transformed non-
linearly.**Form: Y=β0+f1(X1)+f2(X2)+⋯+fn(Xn)+ϵY = \beta_0 + f_1(X_1) +
f_2(X_2) + \cdots + f_n(X_n) + \epsilonY=β0+f1(X1)+f2(X2)+⋯+fn(Xn)+ϵ where
fif_ifi are non-linear functions of predictors, and ϵ\epsilonϵ is the error
term.**Use Case: Useful for modeling complex, non-linear relationships
without assuming a specific functional form for the relationship.
3. Smooth Additive Models:**Description: A type of additive model where
smooth functions are used to model relationships between predictors and the
response.**Form: Similar to GAMs, with smooth functions applied to
predictors but focusing specifically on smoothness in the modeling process.
**Use Case: Suitable for capturing smooth, non-linear effects in the
data.Regression Tree://Purpose: Predicts a continuous target variable by
splitting data into subsets based on feature values.//Process:**Splitting: At
each node, the tree splits the data based on a feature that best separates the
target variable into homogeneous groups.**Leaf Nodes: Each terminal leaf
node represents a predicted value, calculated as the mean of the target
variable in that subset.
Gini Index and Split Criteria:**Gini Index: Primarily used in classification trees,
it measures the impurity of a node. For regression trees, the focus is on
variance reduction rather than Gini Index.**Split Criteria for Regression
Trees:Variance Reduction: Chooses splits that minimize the variance of the
target variable within the resulting subsets. The goal is to reduce the variance
in the target variable as much as possible with each split, improving predictive
accuracy.
Gradient definition:The gradient is a vector that indicates the direction and
rate of the steepest increase of a function. It is used in optimization to find the
minimum or maximum of a function.
Components:*For a function f(x)f(x)f(x) with multiple variables x1,x2,…,xn the
gradient is a vector of partial derivatives: