🎉Community Raffle - Win $25

An exclusive raffle opportunity for active members like you! Complete your profile, answer questions and get your first accepted badge to enter the raffle.
Join and Win

Process failed exception, any clue?

User: "confusedMonMon"
New Altair Community Member
Updated by Jocelyn
I've created a process model that works fine on a sample dataset. However, when I run the process on my whole dataset it gets failed. I'm not sure is it because of the size of the processed files/documents? is there any size limit for the procssed documents in rapidminer? or is it something wrong with the process itself? The exception I'm getting:
  • Exception: java.lang.StackOverflowError
  • Message: null
  • Stack trace:
  • sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
  • sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
  • sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
  • java.lang.reflect.Constructor.newInstance(Constructor.java:423)
  • java.util.concurrent.ForkJoinTask.getThrowableException(ForkJoinTask.java:598)
  • java.util.concurrent.ForkJoinTask.get(ForkJoinTask.java:1005)
  • com.rapidminer.studio.concurrency.internal.AbstractConcurrencyContext.collectResults(AbstractConcurrencyContext.java:206)
  • com.rapidminer.studio.concurrency.internal.StudioConcurrencyContext.collectResults(StudioConcurrencyContext.java:33)
  • com.rapidminer.studio.concurrency.internal.AbstractConcurrencyContext.call(AbstractConcurrencyContext.java:141)
  • com.rapidminer.studio.concurrency.internal.StudioConcurrencyContext.call(StudioConcurrencyContext.java:33)
  • com.rapidminer.Process.executeRootInPool(Process.java:1349)
  • com.rapidminer.Process.execute(Process.java:1314)
  • com.rapidminer.Process.run(Process.java:1291)
  • com.rapidminer.Process.run(Process.java:1177)
  • com.rapidminer.Process.run(Process.java:1130)
  • com.rapidminer.Process.run(Process.java:1125)
  • com.rapidminer.Process.run(Process.java:1115)
  • com.rapidminer.gui.ProcessThread.run(ProcessThread.java:65)
  • Cause
  • Exception: java.lang.StackOverflowError
  • Message: null
  • Stack trace:
  • java.util.regex.Pattern$Branch.match(Pattern.java:4606)
  • java.util.regex.Pattern$GroupHead.match(Pattern.java:4660)
  • java.util.regex.Pattern$LazyLoop.match(Pattern.java:4849)
  • .........................

Thank you

Find more posts tagged with

Sort by:
1 - 1 of 11
    User: "David_A"
    New Altair Community Member
    Accepted Answer
    Hi,

    text processing can be quite memory expensive. Simply keeping the different strings in memory has always some overhead and even filtering with simple regex requires additional memory. So it's really hard to say at which point it will break in your example, all I can say is, that it does not look like a general bug on our side.

    I would suggest to try if you can process the text from one file (the largest one). If this works, your memory in general is sufficient for the task and you can work on reducing the memory footprint of your process.

    For example:
    1) disable the parallel execution of the loop files operator  (with that, not several files are loaded at once)
    2) store some intermediate results (for example all the created documents after "Read Documents") and then do the regex filtering for each document independently

    Best,
    David