N Fold Test
- How do you know if your classifier is any good?
- Obviously, it categorises all the data correctly.
- So, an obvious algorithm is to cache all the training data
away.
- If you run again, classify it the same way. That always works.
- However, you usually want the system to generalise to unseen data.
- Will caching generalise?
- Will the wealth rule?
- Will a line?
- Typically, we test machine learning rules by partitioning the data
into proper subsets.
- For example, we make equal sized training and test sets.
- The first fold trains on the training set, and then tests on
the test set.
- Then, to protect against bad splits, we do it again reversing
the process.
- We throw away the line (or whatever the system is) train on the
test set, and test on the training set.
- That's a 2-Fold test. You can have any fold, e.g 5-Fold.
- Break the data into 5 sets, and run the learning algorithm 5 times.
- Train on 4 and test on 1 (or train on 1 and test on 4).
- Is this generalisable? If the original data is representative, then
it probably is.