首页 > > 详细

辅导 DDA3020、讲解 Python编程设计

DDA3020 Homework 1
Due date: March 09, 2025
Instructions
• The deadline is 23:59, March 09, 2025.
• The weight of this assignment in the final grade is 20%.
• Electronic submission: Turn in solutions electronically via Blackboard. Be sure to submit
your answers as one pdf file plus two python scripts for programming questions. Please
name your solution files as“DDA3020HW1 studentID name.pdf”, “HW1 name Q1.ipynb” and
“HW1 name Q2.ipynb” (“.py” files are also acceptable).
• The complete and executable codes must be submitted. If you only fill in some of the results
in your answer report for programming questions and do not submit the source code (.py or
.ipynb files), you will receive 0 points for the question.
• Note that late submissions will result in discounted scores: 0-48 hours → 50%, more hours
→ 0%.
• Answer the questions in English. Otherwise, you’ll lose half of the points.
• Collaboration policy: You need to solve all questions independently and collaboration between
students is NOT allowed.
1 Written Problems (50 points)
1.1. (Closed-Form Solution, 25 points) Given a ridge regression with data X ∈ R
n×m,
y ∈ R
n
, where n is the number of data, m is the number of attributes used for prediction, under
the assumption that data X is centered, i.e., x¯ = n
1 P n
i=1 xi = 0, show that the closed-form solution
of the cost function
J(w, w0) = (y − Xw − w01)
T
(y − Xw − w01) + λwT w
is
wˆ0 = y¯
w = (XT X + λI)
−1XT y
1
DDA3020 Machine Learning Spring 2025, CUHKSZ
1.2. (Support Vector Machine, 25 points) Consider a dataset with 2 points in 1D: (x1 =
0, y1 = −1) and (x2 = 1, y2 = 1). Consider mapping each point to 3D using the feature vector
ϕ(x) = [1,

2x, x2
]
⊤. (This is equivalent to using a second-order polynomial kernel.) The max
margin classifier has the form:
min ∥w∥
2
s.t.
y1
wT ϕ(x1) + w0
 ≥ 1
y2
wT ϕ(x2) + w0
 ≥ 1
(i) Write down a vector that is parallel to the optimal vector w.
(ii) What is the value of the margin achieved by this w? Hint: recall that the margin is the
distance from each support vector to the decision boundary.
(iii) Solve for w, using the fact that the margin is equal to 1/∥w∥.
(iv) Solve for w0 using your value for w and the optimization problem above. Hint: the points
will be on the decision boundary, so the inequalities will be tight, think about the geometry of 2
points in space, with a line separating one from the other.
(v) Write down the discriminant function f(x) = w0 + w⊤ϕ(x) as an explicit function of x.
2 Programming (50 points)
2.1. (Logistic Regression, 25 points) In this question, we will implement multi-class logistic
regression using gradient descent. Follow code template provided in logistic assignment1.ipynb to
solve this question. Noticably, this question should not be solved by using sklearn package, aiming to
promote your understanding on gradient descent. A mathematical derivation of gradient is required
in this problem. Provide your solution in your written answers.
2.2. (Support Vector Machine, 25 points) In this question, we will explore the use of Support
Vector Machines (SVM) for both linear and non-linear tasks using the sklearn library, as outlined in
the SVM assignment1.ipynb file. By following the instructions in the notebook, we will implement
both types of SVMs to gain a foundational understanding of how to apply SVMs.
2

联系我们
  • QQ:99515681
  • 邮箱:99515681@qq.com
  • 工作时间:8:00-21:00
  • 微信:codinghelp
热点标签

联系我们 - QQ: 99515681 微信:codinghelp
程序辅导网!