Choose a topic to test your knowledge and improve your Machine Learning (ML) skills
What are support vectors?
A perceptron adds up all the weighted inputs it receives, and if it exceeds a certain value, it outputs a 1, otherwise it just outputs a 0.
What is the purpose of the Kernel Trick?
Which of the following can only be used when training data are linearlyseparable?
The firing rate of a neuron
Which of the following evaluation metrics can not be applied in case of logistic regression output to compare with target?
The cost parameter in the SVM means:
The kernel trick
How does the bias-variance decomposition of a ridge regression estimator compare with that of ordinaryleast squares regression?
Which of the following are real world applications of the SVM?
How can SVM be classified?
Which of the following can help to reduce overfitting in an SVM classifier?
Suppose you have trained an SVM with linear decision boundary after training SVM, you correctly infer that your SVM model is under fitting. Which of the following is best option would you more likely to consider iterating SVM next time?
What is/are true about kernel in SVM? 1. Kernel function map low dimensional data to high dimensional space2. Itβs a similarity function
You trained a binary classifier model which gives very high accuracy on the training data, but much lower accuracy on validation data. Which is false.
Suppose your model is demonstrating high variance across the different training sets. Which of the following is NOT valid way to try and reduce the variance?
Suppose you are using RBF kernel in SVM with high Gamma value. What does this signify?
We usually use feature normalization before using the Gaussian kernel in SVM. What is true about feature normalization? 1. We do feature normalization so that new feature will dominate other 2. Some times, feature normalization is not feasible in case of categorical variables 3. Feature normalization always helps when we use Gaussian kernel in SVM
Wrapper methods are hyper-parameter selection methods that
Which of the following methods can not achieve zero training error on any linearly separable dataset?
Suppose we train a hard-margin linear SVM on n > 100 data points in R2, yielding a hyperplane with exactly 2 support vectors. If we add one more data point and retrain the classifier, what is the maximum possible number of support vectors for the new hyperplane (assuming the n + 1 points are linearly separable)?
Let S1 and S2 be the set of support vectors and w1 and w2 be the learnt weight vectors for a linearly separable problem using hard and soft margin linear SVMs respectively. Which of the following are correct?
Which statement about outliers is true?
If TP=9 FP=6 FN=26 TN=70 then Error rate will be
Imagine, you are solving a classification problems with highly imbalanced class. The majority class is observed 99% of times in the training data. Your model has 99% accuracy after taking the predictions on test data. Which of the following is true in such a case? 1. Accuracy metric is not a good idea for imbalanced class problems. 2.Accuracy metric is a good idea for imbalanced class problems. 3.Precision and recall metrics are good for imbalanced class problems. 4.Precision and recall metrics arenβt good for imbalanced class problems.
he minimum time complexity for training an SVM is O(n2). According to this fact, what sizes of datasets are not best suited for SVMβs?
Perceptron Classifier is
Type of dataset available in Supervised Learning is
which among the following is the most appropriate kernel that can be used with SVM to separate the classes
The SVMs are less effective when
What is the precision value for following confusion matrix of binary classification?
Which of the following are components of generalization Error?
Which of the following is not a kernel method in SVM?
During the treatement of cancer patients , the doctor needs to be very careful about which patients need to be given chemotherapy.Which metric should we use in order to decide the patients who should given chemotherapy?
Which one of the following is suitable? 1. When the hypothsis space is richer, overfitting is more likely. 2. when the feature space is larger , overfitting is more likely.
Which of the following is a categorical data?
The soft margin SVM is more preferred than the hard-margin SVM when-
In SVM which has quadratic kernel function of polynomial degree 2 that has slack variable C as one hyper paramenter. What would happen if we use very large value for C
In SVM, RBF kernel with appropriate parameters to perform binary classification where the data is non-linearly seperable. In this scenario
Which of the following is true about SVM? 1. Kernel function map low dimensional data to high dimensional space. 2. It is a similarity Function
What is the Accuracy in percentage based on following confusion matrix of three class classification. Confusion Matrix C=[14 0 0] [ 1 15 0] [ 0 0 6]
Which of the following method is used for multiclass classification?
Based on survey , it was found that the probability that person like to watch serials is 0.25 and the probability that person like to watch netflix series is 0.43. Also the probability that person like to watch serials and netflix sereis is 0.12. what is the probability that a person doesn't like to watch either?
A machine learning problem involves four attributes plus a class. The attributes have 3, 2, 2, and 2 possible values each. The class has 3 possible values. How many maximum possible different examples are there?
MLE estimates are often undesirable because
The difference between the actual Y value and the predicted Y value found using a regression equation is called the
Neural networks
Linear Regression is a _______ machine learning algorithm.
Which of the following methods/methods do we use to find the best fit line for data in Linear Regression?
Which of the following methods do we use to best fit the data in Logistic Regression?