Home
Discussions
Community Q&A
H20 Variable Importance
User36964
Deep Learning model of H20 provides "compute variable importance" choice.
If elected the output of Deep Learning Model list the top ten important attributes. Is there a way to increase this number to top 20 or 100?
Find more posts tagged with
AI Studio
H2O
Deep Learning + Neural Nets
Accepted answers
All comments
varunm1
Hello
@User36964
I am sure that H2O is calculating variable importance for all variables in your dataset, I think its the rapidminer view that is restricting you to see all variables. I can see the top 10 and least 10 variables and their importance. I don't see any option to extend this
@hughesfleming68
any suggestion on this?
Thanks
User36964
I wonder why they limit it. Anyone use this extension need to see the all variable importance.
varunm1
I understand your concern
@User36964
, but in some cases, if the attributes are very high (100's), it is difficult to view all of them. There should have been an option to extract variable importance.
You can use explain predictions operator to see which variables impacted your model predictions. I will use this a lot compared to variable importance. One reason is the limitation of H2O variable importance method (Gedeon based) that extracts importance based on weights only from first two layers of a network, but for large networks, it is not good as the deeper networks can influence your variable importance.
Everything has their own limitations
User36964
As you've stated there should be an option to import all the importances.
The explain predictions operator explain the prediction of each data row. Thus, gathering a general (overall) idea of the important attributes is somewhat challenging. May be attribute frequencies for each prediction can be calculated manually to find the most effective supporting and contradicting attributes.
varunm1
Actually, I do some extra analysis using explain predictions operator. A few more operators additional to explain predictions will provide me with attributes that support and contradict correct predictions as well as wrong predictions. If you are interested you can take a look at below thread for the process of extracting attributes based on outcome (correct or incorrect).
I am trying to work on some feature selection techniques based on this operator, if
@IngoRM
does it earlier it will be available in RM.
https://community.rapidminer.com/discussion/54852/explain-predictions-ranking-attributes-that-supports-and-contradicts-correct-predictions#latest
Thanks
User36964
Thanks,
I'm looking forward to see your solutions
Quick Links
All Categories
Recent Discussions
Activity
Unanswered
日本語 (Japanese)
한국어(Korean)