Machine Learning (ML) MCQ Quiz Hub

Machine Learning (ML) MCQ Set 04

Choose a topic to test your knowledge and improve your Machine Learning (ML) skills

What is gini index?





✅ Correct Answer: 4

Tree/Rule based classification algorithms generate ... rule to perform the classification.





✅ Correct Answer: 1

Decision Tree is





✅ Correct Answer: 3

Which of the following is true about Manhattan distance?





✅ Correct Answer: 1

A company has build a kNN classifier that gets 100% accuracy on training data. When they deployed this model on client side it has been found that the model is not at all accurate. Which of the following thing might gone wrong?Note: Model has successfully deployed and no technical issues are found at client side except the model performance





✅ Correct Answer: 1

Which statement is true about the K-Means algorithm? Select one:





✅ Correct Answer: 3

Which of the following can act as possible termination conditions in K-Means? 1. For a fixed number of iterations. 2. Assignment of observations to clusters does not change between iterations. Except for cases with a bad local minimum. 3. Centroids do not change between successive iterations. 4. Terminate when RSS falls below a threshold.





✅ Correct Answer: 4

Which of the following statement is true about k-NN algorithm? 1) k-NN performs much better if all of the data have the same scale 2) k-NN works well with a small number of input variables (p), but struggles when the number of inputs is very large 3) k-NN makes no assumptions about the functional form of the problem being solved





✅ Correct Answer: 4

In which of the following cases will K-means clustering fail to give good results? 1) Data points with outliers 2) Data points with different densities 3) Data points with nonconvex shapes





✅ Correct Answer: 3

his clustering algorithm terminates when mean values computed for the current iteration of the algorithm are identical to the computed mean values for the previous iteration Select one:





✅ Correct Answer: 1

Which one of the following is the main reason for pruning a Decision Tree?





✅ Correct Answer: 4

You've just finished training a decision tree for spam classification, and it is getting abnormally bad performance on both your training and test sets. You know that your implementation has no bugs, so what could be causing the problem?





✅ Correct Answer: 1

The K-means algorithm:





✅ Correct Answer: 3

Which of the following metrics, do we have for finding dissimilarity between two clusters in hierarchical clustering? 1. Single-link 2. Complete-link 3. Average-link





✅ Correct Answer: 4

In which of the following cases will K-Means clustering fail to give good results?





✅ Correct Answer: 4

Hierarchical clustering is slower than non-hierarchical clustering?





✅ Correct Answer: 1

High entropy means that the partitions in classification are





✅ Correct Answer: 2

Suppose we would like to perform clustering on spatial data such as the geometrical locations of houses. We wish to produce clusters of many different sizes and shapes. Which of the following methods is the most appropriate?





✅ Correct Answer: 2

The main disadvantage of maximum likelihood methods is that they are _____





✅ Correct Answer: 4

The maximum likelihood method can be used to explore relationships among more diverse sequences, conditions that are not well handled by maximum parsimony methods.





✅ Correct Answer: 1

Which Statement is not true statement.





✅ Correct Answer: 3

what is Feature scaling done before applying K-Mean algorithm?





✅ Correct Answer: 1

With Bayes theorem the probability of hypothesis H¾ specified by P(H) ¾ is referred to as





✅ Correct Answer: 2

The probability that a person owns a sports car given that they subscribe to automotive magazine is 40%. We also know that 3% of the adult population subscribes to automotive magazine. The probability of a person owning a sports car given that they don’t subscribe to automotive magazine is 30%. Use this information to compute the probability that a person subscribes to automotive magazine given that they own a sports car





✅ Correct Answer: 4

What is the naïve assumption in a Naïve Bayes Classifier.





✅ Correct Answer: 4

What is the actual number of independent parameters which need to be estimated in P dimensional Gaussian distribution model?





✅ Correct Answer: 4

Give the correct Answer for following statements. 1. It is important to perform feature normalization before using the Gaussian kernel. 2. The maximum value of the Gaussian kernel is 1.





✅ Correct Answer: 3

Consider the following dataset. x,y,z are the features and T is a class(1/0). Classify the test data (0,0,1) as values of x,y,z respectively.





✅ Correct Answer: 1

Which of the following statements about Naive Bayes is incorrect?





✅ Correct Answer: 2

How the entries in the full joint probability distribution can be calculated?





✅ Correct Answer: 2

How many terms are required for building a bayes model?





✅ Correct Answer: 3

Skewness of Normal distribution is ___________





✅ Correct Answer: 3

The correlation coefficient for two real-valued attributes is –0.85. What does this value tell you?





✅ Correct Answer: 3

8 observations are clustered into 3 clusters using K-Means clustering algorithm. After first iteration clusters, C1, C2, C3 has following observations: C1: {(2,2), (4,4), (6,6)} C2: {(0,4), (4,0),(2,5)} C3: {(5,5), (9,9)} What will be the cluster centroids if you want to proceed for second iteration?





✅ Correct Answer: 4

In Naive Bayes equation P(C / X)= (P(X / C) *P(C) ) / P(X) which part considers "likelihood"?





✅ Correct Answer: 1

Which of the following option is / are correct regarding benefits of ensemble model? 1. Better performance 2. Generalized models 3. Better interpretability





✅ Correct Answer: 4

What is back propagation?





✅ Correct Answer: 1

Which of the following is an application of NN (Neural Network)?





✅ Correct Answer: 4

Neural Networks are complex ______________ with many parameters.





✅ Correct Answer: 1

Having multiple perceptrons can actually solve the XOR problem satisfactorily: this is because each perceptron can partition off a linear part of the space itself, and they can then combine their results.





✅ Correct Answer: 3

Which one of the following is not a major strength of the neural network approach?





✅ Correct Answer: 1

The network that involves backward links from output to the input and hidden layers is called





✅ Correct Answer: 3

Which of the following parameters can be tuned for finding good ensemble model in bagging based algorithms? 1. Max number of samples 2. Max features 3. Bootstrapping of samples 4. Bootstrapping of features





✅ Correct Answer: 4

What is back propagation? a) It is another name given to the curvy function in the perceptron b) It is the transmission of error back through the network to adjust the inputs c) It is the transmission of error back through the network to allow weights to be adjusted so that the network can learn d) None of the mentioned





✅ Correct Answer: 3

What is the sequence of the following tasks in a perceptron? Initialize weights of perceptron randomly Go to the next batch of dataset If the prediction does not match the output, change the weights For a sample input, compute an output





✅ Correct Answer: 1

In which neural net architecture, does weight sharing occur?





✅ Correct Answer: 4

Which of the following are correct statement(s) about stacking? 1. A machine learning model is trained on predictions of multiple machine learning models 2. A Logistic regression will definitely work better in the second stage as compared to other classification methods 3. First stage models are trained on full / partial feature space of training data





✅ Correct Answer: 3

The F-test





✅ Correct Answer: 3

What is true about an ensembled classifier? 1. Classifiers that are more “sure” can vote with more conviction 2. Classifiers can be more “sure” about a particular part of the space 3. Most of the times, it performs better than a single classifier





✅ Correct Answer: 4

Which of the following option is / are correct regarding benefits of ensemble model? 1. Better performance 2. Generalized models 3. Better interpretability





✅ Correct Answer: 3

Which of the following can be true for selecting base learners for an ensemble? 1. Different learners can come from same algorithm with different hyper parameters 2. Different learners can come from different algorithms 3. Different learners can come from different training spaces





✅ Correct Answer: 4

If you use an ensemble of different base models, is it necessary to tune the hyper parameters of all base models to improve the ensemble performance?





✅ Correct Answer: 2

Generally, an ensemble method works better, if the individual base models have ____________?Note: Suppose each individual base models have accuracy greater than 50%.





✅ Correct Answer: 1

In an election, N candidates are competing against each other and people are voting for either of the candidates. Voters don’t communicate with each other while casting their votes. Which of the following ensemble method works similar to above-discussed election procedure? Hint: Persons are like base models of ensemble method.





✅ Correct Answer: 1

Suppose there are 25 base classifiers. Each classifier has error rates of e = 0.35. Suppose you are using averaging as ensemble technique. What will be the probabilities that ensemble of above 25 classifiers will make a wrong prediction? Note: All classifiers are independent of each other





✅ Correct Answer: 2