McqMate

Q. |
## Suppose you have fitted a complex regression model on a dataset. Now, you are using Ridge regression with tuning parameter lambda to reduce its complexity. Choose the option(s) below which describes relationship of bias and variance with lambda. |

A. | in case of very large lambda; bias is low, variance is low |

B. | in case of very large lambda; bias is low, variance is high |

C. | in case of very large lambda; bias is high, variance is low |

D. | in case of very large lambda; bias is high, variance is high |

Answer» C. in case of very large lambda; bias is high, variance is low |

731

0

Do you find this helpful?

1

View all MCQs in

Machine Learning (ML)No comments yet

- Suppose you have fitted a complex regression model on a dataset. Now, you are using Ridge regression with tuning parameter lambda to reduce its complexity. Choose the option(s) below which describes relationship of bias and variance with lambda.
- Suppose you have fitted a complex regression model on a dataset. Now, you are using Ridge regression with tuning parameter lambda to reduce its complexity. Choose the option(s) below which describes relationship of bias and variance with lambda.
- What is/are true about ridge regression? 1. When lambda is 0, model works like linear regression model 2. When lambda is 0, model doesn't work like linear regression model 3. When lambda goes to infinity, we get very, very small coefficients approaching 0 4. When lambda goes to infinity, we get very, very large coefficients approaching infinity
- What is/are true about ridge regression? 1. When lambda is 0, model works like linear regression model 2. When lambda is 0, model doesn’t work like linear regression model 3. When lambda goes to infinity, we get very, very small coefficients approaching 0 4. When lambda goes to infinity, we get very, very large coefficients approaching infinity
- Regarding bias and variance, which of the following statements are true? (Here ‘high’ and ‘low’ are relative to the ideal model. (i) Models which overfit are more likely to have high bias (ii) Models which overfit are more likely to have low bias (iii) Models which overfit are more likely to have high variance (iv) Models which overfit are more likely to have low variance
- Regarding bias and variance, which of the following statements are true? (Here ‘high’ and ‘low’ are relative to the ideal model. (i) Models which overfit are more likely to have high bias (ii) Models which overfit are more likely to have low bias (iii) Models which overfit are more likely to have high variance (iv) Models which overfit are more likely to have low variance
- We have been given a dataset with n records in which we have input attribute as x and output attribute as y. Suppose we use a linear regression method to model this data. To test our linear regressor, we split the data in training set and test set randomly. What do you expect will happen with bias and variance as you increase the size of training data?
- We have been given a dataset with n records in which we have input attribute as x and output attribute as y. Suppose we use a linear regression method to model this data. To test our linear regressor, we split the data in training set and test set randomly. What do you expect will happen with bias and variance as you increase the size of training data?
- We have been given a dataset with n records in which we have input attribute as x and output attribute as y. Suppose we use a linear regression method to model this data. To test our linear regressor, we split the data in training set and test set randomly. What do you expect will happen with bias and variance as you increase the size of training data?