POM Assignment 4 GIM123
POM Assignment 4 GIM123
SVM
Question: Answer the following questions.
Question Context: 1 – 2
Suppose you are using a Linear SVM classifier with 2 class classification problem. Now you
have been given the following data in which some points are circled red that are
representing support vectors.
1) If you remove the following any one red points from the data. Does the decision
boundary will change?
A) Yes
B) No
Answer?
2) [True or False] If you remove the non-red circled points from the data, the decision
boundary will change?
A) True
B) False
Answer?
B) How accurately the SVM can predict outcomes for unseen data
Answer?
4) When the C parameter is set to infinite, which of the following holds true?
A) The optimal hyperplane if exists, will be the one that completely separates the data
Answer?
Answer?
6) The minimum time complexity for training an SVM is O(n2). According to this fact, what
sizes of datasets are not best suited for SVM’s?
A) Large datasets
B) Small datasets
Answer?
A) Selection of Kernel
B) Kernel Parameters
Answer?
8) Support vectors are the data points that lie closest to the decision surface.
A) TRUE
B) FALSE
Answer?
9) The SVM’s are less effective when:
Answer?
10) Suppose you are using RBF kernel in SVM with high Gamma value. What does this
signify?
A) The model would consider even far away points from hyperplane for modeling
B) The model would consider only the points close to the hyperplane for modeling
C) The model would not be affected by distance of points from hyperplane for modeling
Answer?
Answer?
12)
Suppose you are building a SVM model on data X. The data X can be error prone which
means that you should not trust any specific data point too much. Now think that you want
to build a SVM model which has quadratic kernel function of polynomial degree 2 that uses
Slack variable C as one of it’s hyper parameter. Based upon that give the answer for
following question.
What would happen when you use very large value of C(C->infinity)?
Note: For small C was also classifying all data points correctly
A) We can still classify data correctly for given setting of hyper parameter C
B) We can not classify data correctly for given setting of hyper parameter C
C) Can’t Say
D) None of these
Answer?
13) What would happen when you use very small C (C~0)?
C) Can’t say
D) None of these
Answer?
14) If I am using all features of my dataset and I achieve 100% accuracy on my training
set, but ~70% on validation set, what should I look out for?
A) Underfitting
C) Overfitting
Answer?
15) Which of the following are real world applications of the SVM?
B) Image Classification
Answer?
Question Context: 16 – 18
Suppose you have trained an SVM with linear decision boundary after training SVM, you
correctly infer that your SVM model is under fitting.
16) Which of the following option would you more likely to consider iterating SVM next
time?
Answer?
17) Suppose you gave the correct answer in previous question. What do you think that is
actually happening?
A) 1 and 2
B) 2 and 3
C) 1 and 4
D) 2 and 4
Answer?
18) In above question suppose you want to change one of it’s(SVM) hyperparameter so
that effect would be same as previous questions i.e model will not under fit?
D) None of these
Answer?
19) We usually use feature normalization before using the Gaussian kernel in SVM. What
is true about feature normalization?
A) 1
B) 1 and 2
C) 1 and 3
D) 2 and 3
Answer?
Suppose you are dealing with 4 class classification problem and you want to train a SVM
model on the data for that you are using One-vs-all method. Now answer the below
questions?
20) How many times we need to train our SVM model in such case?
A) 1
B) 2
C) 3
D) 4
Answer?
21) Suppose you have same distribution of classes in the data. Now, say for training 1
time in one vs all setting the SVM is taking 10 second. How many seconds would it require
to train one-vs-all method end to end?
A) 20
B) 40
C) 60
D) 80
Answer?
22) Suppose your problem has changed now. Now, data has only 2 classes. What would
you think how many times we need to train SVM in such case?
A) 1
B) 2
C) 3
D) 4
Answer?
Question context: 23 – 24
Suppose you are using SVM with linear kernel of polynomial degree 2, Now think that you
have applied this on data and found that it perfectly fit the data that means, Training and
testing accuracy is 100%.
23) Now, think that you increase the complexity(or degree of polynomial of this kernel).
What would you think will happen?
C) Nothing will happen since your model was already 100% accurate
D) None of these
Answer?
24) In the previous question after increasing the complexity you found that training
accuracy was still 100%. According to you what is the reason behind that?
1. Since data is fixed and we are fitting more polynomial term or parameters so the
algorithm starts memorizing everything in the data
2. Since data is fixed and SVM doesn’t need to search in big hypothesis space
A) 1
B) 2
C) 1 and 2
D) None of these
Answer?
25) What is/are true about kernel in SVM?
A) 1
B) 2
C) 1 and 2
D) None of these
Answer?