-
Your help to enhance my RNN-LSTM solution
Hello everyone, In the first place, I would like to thank you for this helpful RapidMiner community platform that helped me a lot in solving many issues in developing this early prediction model. Thus, I am seeking the RapidMiner community's advice regarding my process. Actually, I am trying to develop an early prediction…
-
Can confidence value for prediction be interchangeably be used with the likelihood for an event?
For example: 0.978 confidence (yes) can be implied as the likelihood (chances) of this event being true is 97.8%.
-
I am clicking open visualizations for my data and it is not doing anything.
I have a graph that I can see, but when I click open visualizations, it does not do so.
-
How to see the performance of DBSCAN operator in v10.2
This is because DBSCAN operator in v10.2 is different with the v9 already. So which performance operator can i use to see the performance ?
-
DBSCAN performance evaluation
Hi! I'm facing the following problem. I have to compare the performance of K-Means and DBSCAN on a given dataset. I can easily do it with K-Means: But I can't fit the DBSCAN block in place of the K-Means block. The outputs are different and there is no "Cluster Model" to feed the Performance block. The old version of…
-
Why are DBI using Local Random Seed and Determined Good Start Value same in K-Means
Hi, i'm working on text clustering using K-Means and Singular Value Decomposition (SVD). And i'm using parameter Local Random Seed and Determined Good Start Value to show the different. But the DBI value generated using Local Random Seed is always the same as Determine Good Start Value even though I have tried to enter a…
-
Recommendation System - Prediction of item vs Rating?
Hi Everyone, just wanted to get some help with my recommendation system. I am having trouble in the following areas that i need help or suggestions for. 1) Tuning the Hyper parameters in Item - Item, User - User and Matrix to improve results 2) My models are predicting the rating outcome for the product. How do you flip to…
-
How to choose a better dataset result on performance classification?
here my result, data latih as data sample. i want to compare datasets and choose the better one. all i know that the higher accuracy the better and the lower std deviation the better, but i kinda confuse with this case
-
Error in Binomial Performance Classification Operator
Dears I use deep learning for time series prediction in a binary classification problem. The result sometimes just contains a value for prediction (e.g., just "false"). The binominal classification performance operator has an error in this scenario with this message: "The attribute prediction has 1 different values, must…
-
High Accuracy, low recall and low precision - how to optimise this?
Hi experts, I have a dataset with about 40,000 data and would like to do a classification. I have a binominal label (yes/no). To create the model I take a decision tree. Then I want to apply the created model to a training data set (30,000 data) via the operator Apply model. Overall I have a very high accuracy, of almost…
-
Hi! Please how good is Decision Tree in Regression?
I have used the Decision Tree Regression and other regression models (SVR, LR, ANN, GBT, RFR etc.) on my data, and the former is performing better than all. I also took a new set of data for test, and the decision tree still performed better. But I have read about Decision Trees having overfitting problems, can I keep my…
-
Hi I am handling 80,000 rows, with 200 attributes, strong machine, still Turbo preps hangs...ideas?
Enterprise lisence.
-
Computations for Cluster Distance Performance operator
I am having trouble replicating the computations of the "avg. within cluster distance" metrics produced by the Performance (Cluster Distance Performance) operator. The operator documentation states - "avg._within_centroid_distance: The average within cluster distance is calculated by averaging the distance between the…
-
Result of Performance vector??
Hello, i am pretty new in Rapidminer. I looking for some help. Can somebody help me how to explain the result of Performance vector?? PerformanceVector PerformanceVector:
root_mean_squared_error: 2654164663.713 +/- 1356185105.739 (micro average: 2822093032.240 +/- 0.000)
absolute_error: 2089259686.617 +/- 1016873408.107…
-
how to get performance than k-Means Clustering?
Hi everybody I have a problem. I want to use performance after k-Means Clustering. For this aim I must to use map clustering on labels after clustering and when I run this project I saw an error and I must to changing the number of K, while I am not allowed to change the number of K because I am doing thesis and it not…
-
Is standard deviation reported when the Loop Parameters operator is used?
Hi. I have a CV where, if run on it's own (outside of a Loop Parameters operator) I get average accuracy and the standard deviation. However, when I run my CV within the Loop Parameters operator, changing the Neural Net training cycles, I only get accuracy reported for each of my training cycle settings. Is it possible to…
-
In what order can i use replace missing value and normalization?
Hello! :) I'm trying to build a classification model with RF and I have a data set with many missing values. I tried to use RF without replacing missing values but the performance is not good. For this, I'd like to replace missing values with average. What is the better order to insert the operator? Before or after…
-
How to change metadata especially values of a label attirbute?
Hi I have a label attribute that has some values. When I filter the example set to remove those records including a special value for their labels, the remind example set has that special value in its metadata, and my average precision and recall will be reduced even though my example set has not that value. I have to use…
-
Performance (Binominal Classification) Error: Incompatible Label and Prediction
I have a simple example for demonstrating the effect of varying threshold on error, accuracy, sensitivity, etc. It works fine when I select classification error and the accuracy as metrics. However, the moment I add any of the metrics like precision, recall, AUC, I get the error in the Performance (Binominal…
-
[HELP NEEED] Perfomance Classfication Error
I want to see my performance model, my prediction is multi-classification type (0,1,2) but there is an error when it applies performance(Classification) operators "Performance(classification) cannot handle label." Then I try to use the recommendation help solver. There is another error "label and prediction must be the…
-
Can I optimize a custom performance metric?
Hi, I want to generate a F beta score and change beta according to the weight I want to put on precision or recall. And then I would like to add it to a performance vector so I can optimize it. I know there's a 'Performance to data' operator, but what I would need is the reverse. Is there any way I can do this? EDIT: I…
-
outlier detection taking too long for preprocessing
I have tried cleaning the data to detect anomalies, this is the process I have used, but unfortunately due to the long wait time the process is not returning results. I read the CSV file, then replaced the missing values for attributes days to the funded, funded amount, loan amount, and num lenders total. then I chose a…
-
What does it mean by Support Prediction and Contradict Prediction?
Hi, I'm quite new with the RapidMiner. Currently, I'm testing my dataset with Random Forest, Cross-Validation and Explain Prediction. I'm using the Explain Prediction operator and I'm curious with the outcome results especially on support prediction and contradict prediction. What does it mean with support prediction and…
-
Question for standard deviation in split validation
When using split validation for training and testing, how to make the Rapidminer display the standard deviation for auc, sensitivity, and recall et al. ? (now only mean value was displayed)
-
Performance (cost) sample not behaving as expected
I was looking into the performance (cost) component. It comes with a tutorial. The tutorial applies naive bayes to the Golf dataset using split validation. The outcome should be that 1 of 4 items is misclassified. However, if i run it, all items are misclasified as follows (play -> prediction): yes -> no, no-> yes, yes ->…