operators
nn_here
New Altair Community Member
Hi,
Can you please let me know in rapidminer which operator can be used to pull data from an external source(like websites or so).I tried with the operators extract structured data and search webhose.io,but those didnot help.
Tagged:
0
Answers
-
Have you tried the Webmining extension? there are possibilities to use different http based get requests.1
-
Thank you for providing the help. But sorry i am unable to get it.0
-
https://marketplace.rapidminer.com/UpdateServer/faces/product_details.xhtml?productId=rmx_web have you tried and installed this extension? In This extension are operators which allow you to get data from REST APIs or Websites in case this is what you are looking for, otherwise could you please explain in more detail what you want to get data from?
0 -
Hi, yes i have installed this and tried using the operator 'crawl web'.But the process has failed with the errors as given below.
- Exception: java.lang.NoClassDefFoundError
- Message: org/apache/tika/parser/html/HtmlParser
- Stack trace:
- edu.uci.ics.crawler4j.parser.TikaHtmlParser.(TikaHtmlParser.java:34)
- edu.uci.ics.crawler4j.parser.Parser.(Parser.java:42)
- edu.uci.ics.crawler4j.crawler.CrawlController.(CrawlController.java:85)
- com.rapidminer.operator.web.crawler.CrawlerOperator.doWork(CrawlerOperator.java:269)
- com.rapidminer.operator.Operator.execute(Operator.java:1024)
- com.rapidminer.operator.execution.SimpleUnitExecutor.execute(SimpleUnitExecutor.java:77)
- com.rapidminer.operator.ExecutionUnit$2.run(ExecutionUnit.java:804)
- com.rapidminer.operator.ExecutionUnit$2.run(ExecutionUnit.java:799)
- java.base/java.security.AccessController.doPrivileged(Native Method)
- com.rapidminer.operator.ExecutionUnit.execute(ExecutionUnit.java:799)
- com.rapidminer.operator.OperatorChain.doWork(OperatorChain.java:423)
- com.rapidminer.operator.Operator.execute(Operator.java:1024)
- com.rapidminer.Process.executeRoot(Process.java:1476)
- com.rapidminer.Process.lambda$executeRootInPool$5(Process.java:1452)
- com.rapidminer.studio.concurrency.internal.AbstractConcurrencyContext$AdaptedCallable.exec(AbstractConcurrencyContext.java:362)
- java.base/java.util.concurrent.ForkJoinTask.doExec(Unknown Source)
- java.base/java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(Unknown Source)
- java.base/java.util.concurrent.ForkJoinPool.scan(Unknown Source)
- java.base/java.util.concurrent.ForkJoinPool.runWorker(Unknown Source)
- java.base/java.util.concurrent.ForkJoinWorkerThread.run(Unknown Source)
Thanks in advance.0 -
Could you share the process you tried to use please?0
-
Hi,Kindly find the attached document containing the screenshot of the process that has been run.
Thankyou.0 -
This will be fixed. in the meantime here a workaround which should work: <?xml version="1.0" encoding="UTF-8"?><process version="10.3.000">
<context>
<input/>
<output/>
<macros/>
</context>
<operator activated="true" class="process" compatibility="10.3.000" expanded="true" name="Process">
<parameter key="logverbosity" value="init"/>
<parameter key="random_seed" value="2001"/>
<parameter key="send_mail" value="never"/>
<parameter key="notification_email" value=""/>
<parameter key="process_duration_for_mail" value="30"/>
<parameter key="encoding" value="SYSTEM"/>
<process expanded="true">
<operator activated="true" class="web:crawl_web" compatibility="10.0.000" expanded="true" height="68" name="Crawl Web 5.X" width="90" x="246" y="187">
<parameter key="url" value="https://altair.com/"/>
<list key="crawling_rules"/>
<parameter key="write_pages_into_files" value="true"/>
<parameter key="add_pages_as_attribute" value="false"/>
<parameter key="output_dir" value="C:/"/>
<parameter key="extension" value="txt"/>
<parameter key="max_pages" value="100"/>
<parameter key="max_depth" value="1"/>
<parameter key="domain" value="server"/>
<parameter key="delay" value="1000"/>
<parameter key="max_threads" value="1"/>
<parameter key="max_page_size" value="500"/>
<parameter key="user_agent" value="rapid-miner-crawler"/>
<parameter key="obey_robot_exclusion" value="true"/>
<parameter key="really_ignore_exclusion" value="false"/>
</operator>
<connect from_op="Crawl Web 5.X" from_port="Example Set" to_port="result 1"/>
<portSpacing port="source_input 1" spacing="0"/>
<portSpacing port="sink_result 1" spacing="0"/>
<portSpacing port="sink_result 2" spacing="0"/>
<portSpacing port="sink_result 3" spacing="0"/>
</process>
</operator>
</process>0 -
Hi, Thankyou for the reply given .But Iam afraid it is not working as i am getting the same error. Kindly find the screenshot regarding the same.0
-
Hi,Please find the attached screenshot of the mentioned error.0
-
-
Hi we rolled out Web Mining Extension version 10.0.1 which should fixed the issue0