Find more posts tagged with
Sort by:
1 - 9 of
91

There is a polynomial kernel available in LibSVMLearner.
Hi,
you can find many examples in the "sample" directory of RapidMiner. There should also be one for a general regression setting. For the polynomial LibSVM, you have to set the type to one of the both "SVR" types, select the kernel type "polynomial" and define an appropriate degree and values for C. Which parameter values are appropriate can be evaluated by using one of the parameter optimization operators (please also refer to the sample dir). Here is a simple setup (model is applied on the training data - never do this in real life ;-):
Cheers,
Ingo
you can find many examples in the "sample" directory of RapidMiner. There should also be one for a general regression setting. For the polynomial LibSVM, you have to set the type to one of the both "SVR" types, select the kernel type "polynomial" and define an appropriate degree and values for C. Which parameter values are appropriate can be evaluated by using one of the parameter optimization operators (please also refer to the sample dir). Here is a simple setup (model is applied on the training data - never do this in real life ;-):
However, I would usually prefer an RBF kernel or an (additional) feature construction (for example with YAGGA2) instead but if polynomial works for your data this is of course fine.
<operator name="Root" class="Process" expanded="yes">
<operator name="ExampleSetGenerator" class="ExampleSetGenerator">
<parameter key="number_of_attributes" value="1"/>
<parameter key="target_function" value="one variable non linear"/>
</operator>
<operator name="NoiseGenerator" class="NoiseGenerator">
<list key="noise">
</list>
</operator>
<operator name="LibSVMLearner" class="LibSVMLearner">
<parameter key="C" value="10000.0"/>
<list key="class_weights">
</list>
<parameter key="degree" value="2"/>
<parameter key="keep_example_set" value="true"/>
<parameter key="kernel_type" value="poly"/>
<parameter key="svm_type" value="epsilon-SVR"/>
</operator>
<operator name="ModelApplier" class="ModelApplier">
<list key="application_parameters">
</list>
</operator>
</operator>
Cheers,
Ingo
I got really problems in using this regression type. My approach is to load an example set from a database and produce a prediction model. But I think it tooks too much time (e.g. more than 2 minutes) and the results are not satisfiying.
My operator tree is the following:
My operator tree is the following:
I hope someone can help me to solve these problems or can explain how to calculate these model...
<operator name="Root" class="Process" expanded="yes">
<operator name="DatabaseExampleSource" class="DatabaseExampleSource">
<parameter key="database_system" value="HSQLDB"/>
<parameter key="database_url" value="jdbc:hsqldb:file:SnapshotDB"/>
<parameter key="label_attribute" value="SNAPSHOT"/>
<parameter key="query" value="SELECT SID, SNAPSHOT FROM snapshots"/>
<parameter key="username" value="sa"/>
</operator>
<operator name="LibSVMLearner" class="LibSVMLearner">
<parameter key="C" value="10000.0"/>
<list key="class_weights">
</list>
<parameter key="kernel_type" value="poly"/>
<parameter key="svm_type" value="epsilon-SVR"/>
</operator>
<operator name="ModelWriter" class="ModelWriter">
<parameter key="model_file" value="prediction.mod"/>
</operator>
</operator>
Hi,
if the runtime is too high you could try to reduce the value of "C". If the results are not satisfying, I always would try the RBF kernel with an optimized value for gamma / sigma. This often leads to much better fits. Instead of introducing the non-linearity in the learner, you could also construction additional (polynomial) features before learning and simply apply a linear regression scheme afterwards. This is often faster and leads to understandable models.
Cheers,
Ingo
if the runtime is too high you could try to reduce the value of "C". If the results are not satisfying, I always would try the RBF kernel with an optimized value for gamma / sigma. This often leads to much better fits. Instead of introducing the non-linearity in the learner, you could also construction additional (polynomial) features before learning and simply apply a linear regression scheme afterwards. This is often faster and leads to understandable models.
Cheers,
Ingo
Hi,
here are the basic settings for a RBF SVM:
For the parameter optimization, you could have a look into the sample directory (..._Meta.../...ParameterOptimization.xml).
Cheers,
Ingo
here are the basic settings for a RBF SVM:
<operator name="Root" class="Process" expanded="yes">
<operator name="ExampleSetGenerator" class="ExampleSetGenerator">
<parameter key="attributes_lower_bound" value="-20.0"/>
<parameter key="attributes_upper_bound" value="15.0"/>
<parameter key="number_examples" value="300"/>
<parameter key="number_of_attributes" value="1"/>
<parameter key="target_function" value="one variable non linear"/>
</operator>
<operator name="LibSVMLearner" class="LibSVMLearner">
<parameter key="C" value="2000.0"/>
<list key="class_weights">
</list>
<parameter key="gamma" value="1.0"/>
<parameter key="keep_example_set" value="true"/>
<parameter key="svm_type" value="epsilon-SVR"/>
</operator>
<operator name="ModelApplier" class="ModelApplier">
<list key="application_parameters">
</list>
</operator>
</operator>
For the parameter optimization, you could have a look into the sample directory (..._Meta.../...ParameterOptimization.xml).
Cheers,
Ingo