Github repo for the Course: Stanford Machine Learning (Coursera)
Quiz Needs to be viewed here at the repo (because the image solutions cant be viewed as part of a gist)
True or False | Statement | Explanation |
---|---|---|
False | Adding many new features to the model helps prevent overfitting on the training set. | Adding many new features gives us more expressive models which are able to better fit our training set. If too many new features are added, this can lead to overfitting of the training set. |
False | Introducing regularization to the model always results in equal or better performance on examples not in the training set. | If we introduce too much regularization, we can underfit the training set and this can lead to worse performance even for examples not in the training set. |
False | Introducing regularization to the model always results in equal or better performance on the training set. | If we introduce too much regularization, we can underfit the training set and have worse performance on the training set. |
True | Adding a new feature to the model always results in equal or better performance on the training set | Adding many new features gives us more expressive models which are able to better fit our training set. If too many new features are added, this can lead to overfitting of the training set. |
Answer | Explanation |
---|---|
Adding many new features to the model helps prevent overfitting on the training set. |
True or False | Statement | Explanation |
---|---|---|
False | Using a very large value λ cannot hurt the performance of your hypothesis; the only reason we do not set to be too large is to avoid numerical problems. | Using a very large value of λ can lead to underfitting of the training set. |
False | Because regularization causes J(θ) to no longer be convex, gradient descent may not always converge to the global minimum (when λ > 0, and when using an appropriate learning rate α). | Regularized logistic regression and regularized linear regression are both convex, and thus gradient descent will still converge to the global minimum. |
True | Using too large a value of λ can cause your hypothesis to underfit the data. | A large value of results in a large λ regularization penalty and thus a strong preference for simpler models which can underfit the data. |
False | Because logistic regression outputs values 0 <= h0 <= 1, its range of output values can only be "shrunk" slighly by regularization anyway, so regularization is generally not helpful for it. | None needed |
Answer | Explanation |
---|---|
The hypothesis follows the data points very closely and is highly complicated, indicating that it is overfitting the training set |
Answer | Explanation |
---|---|
The hypothesis does not predict many data points well, and is thus underfitting the training set. |
Hi everyone! I have been reviewing the question 1 many times, and studying again the regularization section, and I can't find the correct answer according to the Coursera revision. Could anyone discuss the solution that was evaluated right by Coursera? I'm really interested to find what's wrong and improve it. I know the honor code but i think that its so important to understand it.