Lecture 4: Logistic Regression
Instructor: Prof. Shuai Huang Industrial and Systems Engineering University of Washington
Regression Instructor: Prof. Shuai Huang Industrial and Systems - - PowerPoint PPT Presentation
Lecture 4: Logistic Regression Instructor: Prof. Shuai Huang Industrial and Systems Engineering University of Washington Extend linear model for classification Need a mathematic transfer function to connect 0 + =1
Instructor: Prof. Shuai Huang Industrial and Systems Engineering University of Washington
π
πΎππ¦π with a binary outcome π§
log
π π 1βπ π = πΎ0 + Οπ=1 π
πΎππ¦π.
depth theoretical investigation
methodologically there is much we can translate from linear regression to logistic regression. Conceptually, it inherits the aura of linear regression model and users can assume a similar degree of confidence of linear regression model onto the logistic regression model
π πΈ = Οπ=1
π
π ππ π§π 1 β π ππ
1βπ§π.
π πΈ = Οπ=1
π
π§π log π ππ + 1 β π§π log 1 β π ππ . This could be further transformed into π πΈ = Οπ=1
π
β log 1 + ππΎ0+Οπ=1
π
πΎππ¦ππ β Οπ=1 π
π§π πΎ0 + Οπ=1
π
πΎππ¦ππ , Then we can have Οπ=1
π
π§π log π ππ + 1 β π§π log 1 β π ππ , = Οπ=1
π
log 1 β π ππ β Οπ=1
π
π§π log
π ππ 1βπ ππ ,
= Οπ=1
π
β log 1 + ππΎ0+Οπ=1
π
πΎππ¦ππ β Οπ=1 π
π§π πΎ0 + Οπ=1
π
πΎππ¦ππ .
following formula: πΈπππ₯ = πΈπππ β
π2π πΈ ππΈππΈπ β1 ππ πΈ ππΈ .
ππ πΈ ππΈ = Οπ=1 π
ππ π§π β π ππ ,
π2π πΈ ππΈππΈπ = β Οπ=1 π
ππππ
ππ ππ
1 β π ππ .
ππ πΈ ππΈ = ππ π β π , π2π πΈ ππΈππΈπ = βππππ.
where π is the π Γ π + 1 input matrix, π is the π Γ 1 column vector of π§π, π is the π Γ 1 column vector of π ππ , and π is a π Γ π diagonal matrix of weights with the nth diagonal element as π ππ 1 β π ππ .
Plugging this into the updating formula of the Newton-Raphson algorithm, πΈπππ₯ = πΈπππ β
π2π πΈ ππΈππΈπ β1 ππ πΈ ππΈ ,
we can derive that πΈπππ₯ = πΈπππ + ππππ
β1πππ π β π ,
= ππππ
β1πππ ππΈπππ + πβ1 π β π
, = ππππ
β1ππππ,
where π΄ = ππΈπππ + πβ1 π β π .
regression model, where each data point ππ, π§π is associated with a weight π₯π to reduce the influence of potential outliers in fitting the regression model. πΈπππ₯ β΅ arg min
πΈ
π β ππΈ ππ π β ππΈ .
Least Square or IRLS algorithm. π is referred as the adjusted response.
this?
Putting all these together, a complete flow of the IRLS is shown in below:
1 1+πβ πΎ0+Οπ=1
π πΎππ¦ππ for π = 1,2, β¦ , π.
π ππ 1 β π ππ for π = 1,2, β¦ , π.
β1ππππ.