Leave One Out results in AUC of 0.5

erik_van_ingen
erik_van_ingen New Altair Community Member
edited November 2024 in Community Q&A
My target label is binominal, number of examples is 553. Running supervised classification deep learning with cross validation:
  • 10-fold  results in AUC = 0.846 and Accuracy = 76%
  • Leave 3 Out (180 fold) results in AUC = 0.5 and Accuracy = 42%
How should I interpret this? Which config should I trust? 






Tagged:

Best Answers

  • varunm1
    varunm1 New Altair Community Member
    Answer ✓
    Hello @erik_van_ingen

    Would like to point two things. First, as the data set is small the higher number of folds give lesser test data which means you will have a lot of variance in your results. This is due to the inability of the test set to capture all the underlying distributions in data. I recommend going with a 3 or 5 fold for this dataset.

    Second, I am not sure if you applied any feature selection techniques (forward, backward etc.) on your data, you can do that and see attributes that are helpful in predicting. This might improve your performances and reduce computational complexity as well.

    Thanks for your understanding.

Answers

  • erik_van_ingen
    erik_van_ingen New Altair Community Member
    Classes are imbalanced 467/84 for No/Yes. I am using stratified sampling. 

    Furthermore, I used the generate weight operation to anticipate on the class imbalance. I tested this both outside the cross validation as within the cross validation. 

    I tested other ML operators as well as Naive Bayes, Gradient Boost and so forth. DL performed usually the best. 

    Yes, I am aware that the sample size is relatively low. What ML operator would best fit, given the sample size?







  • varunm1
    varunm1 New Altair Community Member
    Answer ✓
    Hello @erik_van_ingen

    Would like to point two things. First, as the data set is small the higher number of folds give lesser test data which means you will have a lot of variance in your results. This is due to the inability of the test set to capture all the underlying distributions in data. I recommend going with a 3 or 5 fold for this dataset.

    Second, I am not sure if you applied any feature selection techniques (forward, backward etc.) on your data, you can do that and see attributes that are helpful in predicting. This might improve your performances and reduce computational complexity as well.

    Thanks for your understanding.
  • asem_k
    asem_k New Altair Community Member
    Hi there..
    I still don't quite get it why AUC is close to chance whenever Leave-One-Out cross-validation is used.
    I can see why accuracy measure has a high standard-deviation (each fold, you are either getting 100% correct prediction, or 0 percent correct prediction), but how is that also affecting AUC? Is it because how AUC is actually calculated (can you elaborate on this)?
    By the way, class imbalance, modling techniques, and data size, seem not to have effect on this (try it with many variations of above in rapidminer) and same thing is observed about AUC.