Chapter 4: Linear Models For Classification: Grit Hein & Susanne Leiberg
Chapter 4: Linear Models For Classification: Grit Hein & Susanne Leiberg
Chapter 4: Linear Models For Classification: Grit Hein & Susanne Leiberg
weight vector
linear decision
boundary
diameter
• minimize risk (e.g. classifying sick person as healthy more costly than
classifying healthy person as sick)
Discriminant
function
• directly
map each x onto
a class label
Tools
• Least Square
Classification
• Fisher’s Linear
Discriminant
Discriminant functions
Decision region 1
decision boundary
Decision region 2
w determines orientation
of decision boundary
feature 1
ω0 determines location
of decision boundary
Discriminant functions - How to determine
parameters?
• Problem: sensitive to outliers; also distance between the outliers and the
discriminant function is minimized --> can shift function in a way that
leads to misclassifications
least squares
logistic
regression
Discriminant functions - How to determine
parameters?
• you model the posterior probabilities directly assuming that they have a
sigmoid-shaped distribution (without modeling class priors and class-
conditional densities)
p(C2/ϕ) = 1-p(C1/ϕ)
Probabilistic Discriminative Models - Logistic
Regression
• that means starting from some initial values the weights are changed until
the likelihood is maximized
Normalizing posterior probabilities
Z = unknown
p(z) = 1/Z f(z)
normalization constant
• Goal is to find Gaussian
• approximation q(z) centered on
q(z) p(z)
weight vector
linear decision
boundary
diameter
Terminology
• Two classes
• K > 2 classes
• t = (0,1,0,0,0)T