Machine Learning MCQs-3 (Logistic Regression, KNN, SVM, Decision Tree)
Machine Learning MCQs-3
(Logistic Regression, KNN, SVM, Decision Tree)
---------------------------------------------------------------------
1.A Support Vector Machine can be used for
- Performing linear or nonlinear classification
- Performing regression
- For outlier detection
- All of the above
2. The decision boundaries in a Support Vector machine is fully determined (or “supported”) by the instances located on the edge of the street?
- True
- False
3. Support Vector Machines are not sensitive to feature scaling
- True
- False
Ans: 2
4. If we strictly impose that all instances be off the street and on the right side, this is called
- Soft margin classification
- Hard margin classification
- Strict margin classification
- Loose margin classification
Ans: 2
5.The main issues with hard margin classification are
- It only works if the data is linearly separable
- It is quite sensitive to outliers
- It is impossible to find a margin if the data is not linearly separable
- All of the above
Ans: 4
6. The objectives of Soft Margin Classification are to find a good balance between
- Keeping the street as large as possible
- Limiting the margin violations
- Both of the above
Ans: 3
7. The balance between keeping the street as large as possible and limiting margin violations is controlled by this hyperparameter
- tol
- loss
- penalty
- C
Ans: 4
8. A smaller C value leads to a wider street but more margin violations.
- True
- False
Ans: 1
9. If your SVM model is overfitting, you can try regularizing it by reducing the value of
- tol
- C hyperparameter
- intercept_scaling
- None of the above
Ans: 2
10. A similarity function like Gaussian Radial Basis Function is used to
- Measure how many features are related to each other
- Find the most important features
- Find the relationship between different features
- Measure how much each instance resembles a particular landmark
Ans: 4
11. When using SVMs we can apply an almost miraculous mathematical technique for adding polynomial features and similarity features called the
- Kernel trick
- Shell trick
- Mapping and Reducing
- None of the Above
Ans: 1
12. Which is right for the gamma parameter of SVC which acts as a regularization hyperparameter
- If model is overfitting, increase it, if it is underfitting, reduce it
- If model is overfitting, reduce it, if it is underfitting, increase it
- If model is overfitting, keep it same
- If it is underfitting, keep it same
Ans: 2
13. LinearSVC is much faster than SVC(kernel="linear"))
- True
- False
Ans: 1
14. In SVM regression the model tries to
- Fit the largest possible street between two classes while limiting margin violations
- Fit as many instances as possible on the street while limiting margin violations
Ans: 2
15. Decision Trees can be used for
- Classification Tasks
- Regression Tasks
- Multi-output tasks
- All of the above
Ans: 4
16. The iris dataset has
- 5 features and 3 classes
- 4 features and 3 classes
- 2 features and 3 classes
- 4 features and 2 classes
Ans: 2
17. A node’s gini attribute measures
- The number of training instances in the node
- The ratio of training instances in the node
- Its impurity
Ans: 3
18. If all the training instances of a node belong to the same class then the value of the node's Gini attribute will be
- 1
- 0
- Any integer between 0 and 1
- A negative value
Ans: 2
19. A Gini coefficient of 1 expresses maximal inequality among the training samples
- True
- False
Ans: 1
20. Gini index for a node is found by subtracting the sum of the square of ratio of each classes in a node from 1
- True
- False
Ans: 1
21. A decision tree estimates the probability that an instance belongs to a particular class k by finding the corresponding leaf node for the instance and then returning the ratio of training instances of class k
- True
- False
Ans: 1
22. The Decision Tree classifier predicts the class which has the highest probability
- True
- False
Ans: 1
23. The CART algorithm splits the training set in two subsets
- Using all the features and a threshold tk
- Using a single feature k and a threshold tk
- Using half of the features and a threshold k
Ans: 2
24. How does the CART algorithm chooses the feature k and the threshold tk for splitting ?
- It randomly chooses a feature k
- It chooses the mean of the values of the feature k as threshold
- It chooses the feature k and threshold tk which produces the purest subsets
- It chooses the feature k and threshold tk such that the gini index value of the subsets is 0
Ans: 3
25. The cost function for finding the value of feature k and threshold tk takes into consideration
- The Gini index values of the subsets
- The number of instances in the subsets
- The total number of instances in the node that is being split
- All of these
Ans: 4
26. Once the CART algorithm has successfully split the training set in two
- It stops splitting further
- It splits the subsets using the same logic, then the sub- subsets and so on, recursively
- It splits only the right subset
- It splits only the left subset
Ans: 2
27.The CART algorithm stops recursion once it reaches the maximum depth (defined by the max_depth hyperparameter), or if it cannot find a split that will reduce impurity
- True
- False
Ans: 1
28. Which of the following are correct for the CART algorithm
- It is a greedy algorithm
- It greedily searches for an optimum split at each level
- It does not check whether or not the split will lead to the lowest possible impurity several levels down
- All of the above are correct
Ans: 4
29. While making a prediction in Decision Tree, each node only requires checking the value of one feature
- True
- False
Ans: 1
30. Gini impurity is slightly faster to compute in comparison to entropy
- True
- False
Ans: 1
31. Models like Decision Tree models are often called nonparametric model because
- They do not have any parameters
- The number of parameters is not determined prior to training
- They have lesser parameters as compared to other models
- They are easy to interpret and understand
Ans: 2
Comments
Post a Comment