# Learning Curves

- Curve to check the learning algorithm, whether bias, variance, or both

- To plot learning curve, take small subset of training set(10-40 training examples)
- training set for m = 3, training set = 0. slightly larger if using regularization
- Learning curves will only make a shape only based on that three training example
- Say m >= 4, will become harder for the curve to shape all training set perfectly
- Concretely more training examples, the harder the curve to fit all the examples.
- The more data we have, the more fit the cross validation error, that is better hyphothesis for more training examples
- Sooo, with m start small, the error of Jtrain will be small, because it will overfit smaller data

- This is what we get for the curve if what we have is either bias or variance
- First choosing the model (See hypothesis formula) that is only the linear model.
- The high bias can be shown as the three Jtest, Jtrain, and Jcv are all end up at high value
- suffering High Bias can't really help much if we get more training data.
- As we taught earlier, the Jcv will be similar to Jtrain if the learning suffer high bias, the error will be usually high.

- Next is if we suffering high variance, denotes by Jc) >>(much higher) Jtrain.
- Suppose we use hypothesis with a lot of high order polynomials and small lambda
- Jtrain if get more data, will generally close to future example,
- But more the data the cross validation will keep decreasing(less error), and the Jtrain will keep increasing(more error)
- Sometimes curve can be a lot cleaner/messier, but still helps out intuition if it has bias/variance problem

- Next, discussed about spesific actions to take (or not take) to increase the performance of learning algorithm

Contents © 2020 Jonathan Hari Napitupulu. All code except stated otherwise is protected by MIT License - Powered by Nikola