Assignment 3: Introduction To Machine Learning Prof. B. Ravindran
Assignment 3: Introduction To Machine Learning Prof. B. Ravindran
1
4. Which of the following is correct about linear discriminant analysis ?
(a) It minimizes the variance between the classes relative to the within class variance.
(b) It maximizes the within class variance relative to the variance between classes.
(c) It maximizes the variance between the classes relative to the within class variance.
(d) None of these
Sol. (c)
Refer to the videos
5. (2 marks) Consider the case where two classes follow Gaussian distribution which are cen-
tered at (−1, 2) and (1, 4) and have identity covariance matrix. Which of the following is the
separating decision boundary?
(a) y − x = 3
(b) x + y = 3
(c) x + y = 6
(d) (b) and (c) are possible
(e) None of these
(f) Can not be found from the given information
Sol. (b)
As the distribution is Gaussian and have identity covariance (which are equal), the separating
boundary will be linear. The decision boundary will be orthogonal to the line joining the
centers and will pass from the midpoint of centers.
6. Consider the following data with two classes. The color indicates different class.
Which of the following models (with NO additional complexity) can achieve zero training error
for classification?
(a) LDA
(b) PCA
2
(c) Logistic regression
(d) None of these
Sol. (d)
All the classifiers in option (a), (b) and (c) are linear. The training data is linearly non-
separable.
7. Which of the following technique is well known to utilize class labels in feature selection for
classification?
(a) LDA
(b) PCA (by dimensionality reduction)
(c) both (a) and (b)
(d) None of these
Sol. (a)
PCA does not use the class labels. It chooses the directions with maximum variance in the
data.
8. We discussed the use of MLE for the estimation of parameters of logistic regression model. We
used which of the following assumptions to derive the likelihood function ?
(a) independence among the class labels
(b) independence among each training sample
(c) independence among the parameters of the model
(d) None of these
Sol. (b)
9. Consider the following distribution of training data:
3
(a) Linear Discriminant Analysis
(b) Principal Component Analysis
(c) (a) and (b) perform very poorly, so have to choose Quadratic Discriminant Analysis
(d) (a) or (b) are equally good
(e) None of these
Sol. (d)
The direction of maximum variance is along the direction X1 = X2 . The projected points in
this direction can be easily classified into two classes correctly.
LDA can find a linearly separable decision boundary as the data is linearly spearable.