Why is the Confusion matrix generated by Performance Vector and the Predictive model is different?
I am using Gradient Boosted Trees for my dataset. The process output shows the Performance Vector which gives the accuracy and confusion matrix, and Gradient Boosted Model which gives the model metrics, Confusion matrix, Variable importance, model summary and scoring history. But both the confusion matrix are different. Which should confusion matrix should I consider to evaluate my model?
Find more posts tagged with
Sort by:
1 - 1 of
11
Hi,
the confusion matrix in the model are _training_ errors. So you should usually work on the Performance Vector, not the Gradient Boosted Model values. These are sometimes interesting to have a look on overfitting (e.g. add more complexity or not).
~Martin
Hi,
the confusion matrix in the model are _training_ errors. So you should usually work on the Performance Vector, not the Gradient Boosted Model values. These are sometimes interesting to have a look on overfitting (e.g. add more complexity or not).
~Martin