RM 9.1 feedback : Auto-model documentation

lionelderkrikor
lionelderkrikor New Altair Community Member
edited November 5 in Community Q&A
Hi,

I see that cross validation is now used to evaluate the performance of models in Auto-Model.
I see that the performance associated to the optimized model (calculated via a  3 - folds CV on the whole training set -by defaut 60% of the dataset- )  is different of the performance of the model delivered by the Performance average (Robust) operator (calculated via a  (7-2 = 5 -by default- folds)CV on the test set - 40 % of the dataset). I think that this principe of evaluation of the performances must be explained in the documentation of Auto-Model (in the documentation of the "results" screen). Moreover the actual documentation is out of date : 


Generally, I think that these elements are important and must be read and understood by the user.

I have a subsidiary question about Auto-Model :
Why the data sampling is different according th the used model, for example :  
 NB ==> max 2000000 examples
 SVM ==> max 10000 examples ?

Thank you for your attention,
Regards,

Lionel
  

Best Answers

  • IngoRM
    IngoRM New Altair Community Member
    Answer ✓
    Hi Lionel,
    You are right - we will update the documentation accordingly.  Sorry for this oversight.
    Please note that the outer validation is not a full cross-validation but a multiple hold-out set approach with a robust average calculation (by removing the two outliers).  While this estimation is obviously not as good as a full-blown cross-validation, it comes close plus it has a lower runtime and delivers at least some idea of the deviation of the results.
    The different sample sizes are used to ensure an acceptable runtime for the complete AM run.  The different algorithms have all different algorithmic complexities.  Naive Bayes for example can be calculated with a single data scan (i.e. linear runtime which is as fast as it gets).  An SVM on the other hand has a cubic runtime which would take ages on millions of data rows.
    Best,
    Ingo
  • IngoRM
    IngoRM New Altair Community Member
    Answer ✓
    Just FYI: the documentation has been updated and will be delivered with the next release.
    Thanks again for pointing this out,
    Ingo

Answers

  • IngoRM
    IngoRM New Altair Community Member
    Answer ✓
    Hi Lionel,
    You are right - we will update the documentation accordingly.  Sorry for this oversight.
    Please note that the outer validation is not a full cross-validation but a multiple hold-out set approach with a robust average calculation (by removing the two outliers).  While this estimation is obviously not as good as a full-blown cross-validation, it comes close plus it has a lower runtime and delivers at least some idea of the deviation of the results.
    The different sample sizes are used to ensure an acceptable runtime for the complete AM run.  The different algorithms have all different algorithmic complexities.  Naive Bayes for example can be calculated with a single data scan (i.e. linear runtime which is as fast as it gets).  An SVM on the other hand has a cubic runtime which would take ages on millions of data rows.
    Best,
    Ingo
  • lionelderkrikor
    lionelderkrikor New Altair Community Member
    Hi @IngoRM,

    Thanks for your detailed answer . It's clear in my mind now.
    Regards,

    Lionel
  • IngoRM
    IngoRM New Altair Community Member
    Answer ✓
    Just FYI: the documentation has been updated and will be delivered with the next release.
    Thanks again for pointing this out,
    Ingo