TOPICS (Click to Navigate)

Pages

Sunday, January 31, 2021

Machine Learning Multiple Choice Questions and Answers 25

Top 3 Machine Learning Quiz Questions with Answers explanation, Interview questions on machine learning, quiz questions for data scientist answers explained, machine learning exam questions, question bank in machine learning, overfitting in non-parametric machine learning algorithms, decision tree, linear regression


Machine learning Quiz Questions - Set 25

 

1. Which of the following is a disadvantage of non-parametric machine learning algorithms?

a) Capable of fitting a large number of functional forms (Flexibility)

b) Very fast to learn (Speed)

c) More of a risk to overfit the training data (Overfitting)

d) They do not require much training data

Answer: (c) More of a risk to overfit the training data

Overfitting happens when a model learns the detail and noise in the training data to the extent that it negatively impacts the performance of the model on new data. Overfitting is more likely with nonparametric and nonlinear models that have more flexibility when learning a target function. For example, decision trees are a nonparametric machine learning algorithm that is very flexible and is subject to overfitting training data.

[For more, please refer Overfitting and Underfitting With Machine Learning Algorithms]

 

2. A decision tree has low training error and a large test error. What is the possible problem?

a) Decision tree is too shallow

b) Learning rate too high

c) There is too much training data

d) Decision tree is overfitting

Answer: (d) Decision tree is overfitting

Overfitting causes low training error. Overfitting means that the model predicts the (training) data too well. It is too good to be true. If the new data point comes in, the prediction may be wrong.

Pruning can help in reducing the complexity of the final classifier, and hence improves predictive accuracy by the reduction of overfitting.

 

3. Suppose we have a regularized linear regression model. What is the effect of increasing λ on bias and variance?

a) Increases bias, increases variance

b) Increases bias, decreases variance

c) Decreases bias, increases variance

d) Decreases bias, decreases variance

Answer: (b) Increases bias, decreases variance

Increasing λ increases bias and decreases variance

Regularized regression

It is a type of regression where the coefficient estimates are constrained to zero. The magnitude (size) of coefficients, as well as the magnitude of the error term are penalized. Complex models are discouraged, primarily to avoid overfitting. In other words, this technique discourages learning a more complex or flexible model, so as to avoid the risk of overfitting. [For more refer here – regularized regression, ] and [Refer here - regularization ]

Type of regularized regression

Ridge regression (L2 regularization)

Lasso regression (L1 regularization)

 

**********************

Related links:

List the type of regularized regression

Multiple choice quiz questions in machine learning

What is regularized linear regression 

low training error vs large test error in decision tree

What is the disadvantage of non-parametric machine learning algorithms

No comments:

Post a Comment