K-fold cross validation
. To
understand it a bit better, imagine 5-fold cross validation, where we move the testing set by
20 since there are 5 rows. Then we move the remaining set with the dataset and find the
average of all the folds:
Quite confusing, right? But scikit-learn has built-in support for cross validation. This
feature will be a good way to make sure that we are not overfitting our model and we are
not running our model on a bad testing set.
Decision trees
In this section, we will be using decision trees and student performance data to predict
whether a child will do well in school. We will use the previous techniques with some
scikit-learn code. Before starting with the prediction, let's just learn a bit about what
decision trees are.
Building Your Own Prediction Models
Chapter 1
Do'stlaringiz bilan baham: |