联系方式

  • QQ:99515681
  • 邮箱:99515681@qq.com
  • 工作时间:8:00-23:00
  • 微信:codinghelp

您当前位置:首页 >> Python编程Python编程

日期:2025-03-12 10:15

DDA3020 Homework 1

Due date: March 09, 2025

Instructions

• The deadline is 23:59, March 09, 2025.

• The weight of this assignment in the final grade is 20%.

• Electronic submission: Turn in solutions electronically via Blackboard. Be sure to submit

your answers as one pdf file plus two python scripts for programming questions. Please

name your solution files as“DDA3020HW1 studentID name.pdf”, “HW1 name Q1.ipynb” and

“HW1 name Q2.ipynb” (“.py” files are also acceptable).

• The complete and executable codes must be submitted. If you only fill in some of the results

in your answer report for programming questions and do not submit the source code (.py or

.ipynb files), you will receive 0 points for the question.

• Note that late submissions will result in discounted scores: 0-48 hours → 50%, more hours

→ 0%.

• Answer the questions in English. Otherwise, you’ll lose half of the points.

• Collaboration policy: You need to solve all questions independently and collaboration between

students is NOT allowed.

1 Written Problems (50 points)

1.1. (Closed-Form Solution, 25 points) Given a ridge regression with data X ∈ R

n×m,

y ∈ R

n

, where n is the number of data, m is the number of attributes used for prediction, under

the assumption that data X is centered, i.e., x¯ = n

1 P n

i=1 xi = 0, show that the closed-form solution

of the cost function

J(w, w0) = (y − Xw − w01)

T

(y − Xw − w01) + λwT w

is

wˆ0 = y¯

w = (XT X + λI)

−1XT y

1

DDA3020 Machine Learning Spring 2025, CUHKSZ

1.2. (Support Vector Machine, 25 points) Consider a dataset with 2 points in 1D: (x1 =

0, y1 = −1) and (x2 = 1, y2 = 1). Consider mapping each point to 3D using the feature vector

ϕ(x) = [1,

2x, x2

]

⊤. (This is equivalent to using a second-order polynomial kernel.) The max

margin classifier has the form:

min ∥w∥

2

s.t.

y1

 wT ϕ(x1) + w0

 ≥ 1

y2

 wT ϕ(x2) + w0

 ≥ 1

(i) Write down a vector that is parallel to the optimal vector w.

(ii) What is the value of the margin achieved by this w? Hint: recall that the margin is the

distance from each support vector to the decision boundary.

(iii) Solve for w, using the fact that the margin is equal to 1/∥w∥.

(iv) Solve for w0 using your value for w and the optimization problem above. Hint: the points

will be on the decision boundary, so the inequalities will be tight, think about the geometry of 2

points in space, with a line separating one from the other.

(v) Write down the discriminant function f(x) = w0 + w⊤ϕ(x) as an explicit function of x.

2 Programming (50 points)

2.1. (Logistic Regression, 25 points) In this question, we will implement multi-class logistic

regression using gradient descent. Follow code template provided in logistic assignment1.ipynb to

solve this question. Noticably, this question should not be solved by using sklearn package, aiming to

promote your understanding on gradient descent. A mathematical derivation of gradient is required

in this problem. Provide your solution in your written answers.

2.2. (Support Vector Machine, 25 points) In this question, we will explore the use of Support

Vector Machines (SVM) for both linear and non-linear tasks using the sklearn library, as outlined in

the SVM assignment1.ipynb file. By following the instructions in the notebook, we will implement

both types of SVMs to gain a foundational understanding of how to apply SVMs.

2


版权所有:留学生编程辅导网 2020 All Rights Reserved 联系方式:QQ:99515681 微信:codinghelp 电子信箱:99515681@qq.com
免责声明:本站部分内容从网络整理而来,只供参考!如有版权问题可联系本站删除。 站长地图

python代写
微信客服:codinghelp