"Erratic behaviour of some learning operators with the Text Plugin"
Legacy User
New Altair Community Member
Hello,
I get strange behaviour with some learning operators using the Text plugin.
For example, using the Text Plugin Sample 01_TextClassificationXVal.xml, when replacing the learning operator by NaiveBayes, all samples go to the same class. Using the Weka version of NaiveBayes, it works fine.
With the same example, when replacing the learning operator by LibSVMLearner, using the default parameter C-SVC and rbf, the results are completely inverted (no true positive, no true negative). With other parameters, there are no problems.
It is not clear whether this problem come from the Text Plugin or Rapidminer itself.
Does anyone have an explanation?
Thanks
I get strange behaviour with some learning operators using the Text plugin.
For example, using the Text Plugin Sample 01_TextClassificationXVal.xml, when replacing the learning operator by NaiveBayes, all samples go to the same class. Using the Weka version of NaiveBayes, it works fine.
With the same example, when replacing the learning operator by LibSVMLearner, using the default parameter C-SVC and rbf, the results are completely inverted (no true positive, no true negative). With other parameters, there are no problems.
It is not clear whether this problem come from the Text Plugin or Rapidminer itself.
Does anyone have an explanation?
Thanks
Tagged:
0
Answers
-
Hi,
NaiveBayes tries to estimate the normal distribution of the data. If all attribute values are 0, which is the case within this sample process, the resulting normal distribution has mean 0 and variance 0, causing an infinite density at this value and all other values having density 0. Since NaiveBayes assumes independence between all attributes, the probabilities are multiplied. With one empty attribute, the product will become 0, causing the classification to only one class.
The svm seems to invert mappings.
That's for explantion. And now we will start working on it Thanks for the hint.
Greeting,
Sebastian0