Constannnnnt / Distributed-CoreNLP

This infrastructure, built on Stanford CoreNLP, MapReduce and Spark with Java, aims at processing documents annotations at large scale.
https://github.com/Constannnnnt/Distributed-CoreNLP
MIT License
0 stars 0 forks source link

OOM Error when running NATLOG and OPENIE On SimpleNLP (dependency parser) #2

Closed ji-xin closed 5 years ago

ji-xin commented 5 years ago

An error that is likely to be caused by the dependency parser.

2018-11-12 16:46:10 INFO  MaxentTagger:88 - Loading POS tagger from edu/stanford/nlp/models/pos-tagger/english-left3words/english-left3words-distsim.tagger ... done [0.5 sec].
2018-11-12 16:46:10 INFO  DependencyParser:88 - Loading depparse model: edu/stanford/nlp/models/parser/nndep/english_UD.gz ... 
2018-11-12 16:46:16 ERROR Utils:91 - Aborting task
java.lang.OutOfMemoryError: Java heap space
        at edu.stanford.nlp.parser.nndep.Classifier.preCompute(Classifier.java:662)
        at edu.stanford.nlp.parser.nndep.Classifier.preCompute(Classifier.java:644)
        at edu.stanford.nlp.parser.nndep.DependencyParser.initialize(DependencyParser.java:1189)
        at edu.stanford.nlp.parser.nndep.DependencyParser.loadModelFile(DependencyParser.java:630)
        at edu.stanford.nlp.parser.nndep.DependencyParser.loadFromModelFile(DependencyParser.java:499)
        at edu.stanford.nlp.pipeline.DependencyParseAnnotator.<init>(DependencyParseAnnotator.java:57)
        at edu.stanford.nlp.pipeline.AnnotatorImplementations.dependencies(AnnotatorImplementations.java:240)
        at edu.stanford.nlp.simple.Document$5.lambda$null$0(Document.java:147)
        at edu.stanford.nlp.simple.Document$5$$Lambda$63/1641616399.get(Unknown Source)
        at edu.stanford.nlp.util.Lazy$2.compute(Lazy.java:106)
        at edu.stanford.nlp.util.Lazy.get(Lazy.java:31)
        at edu.stanford.nlp.simple.Document$5.get(Document.java:147)
        at edu.stanford.nlp.simple.Document$5.get(Document.java:142)
        at edu.stanford.nlp.simple.Document.runDepparse(Document.java:918)
        at edu.stanford.nlp.simple.Document.runNatlog(Document.java:938)
        at edu.stanford.nlp.simple.Sentence.natlogPolarities(Sentence.java:897)
        at edu.stanford.nlp.simple.Sentence.natlogPolarities(Sentence.java:905)
        at ca.uwaterloo.cs651.project.SimpleNLP.lambda$main$71c840f9$1(SimpleNLP.java:114)
        at ca.uwaterloo.cs651.project.SimpleNLP$$Lambda$27/273072814.call(Unknown Source)
        at org.apache.spark.api.java.JavaPairRDD$$anonfun$toScalaFunction$1.apply(JavaPairRDD.scala:1040)
        at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
        at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
        at org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$4.apply(SparkHadoopWriter.scala:125)
        at org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$4.apply(SparkHadoopWriter.scala:123)
        at org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1414)
        at org.apache.spark.internal.io.SparkHadoopWriter$.org$apache$spark$internal$io$SparkHadoopWriter$$executeTask(SparkHadoopWriter.scala:135)
        at org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$3.apply(SparkHadoopWriter.scala:79)
        at org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$3.apply(SparkHadoopWriter.scala:78)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
        at org.apache.spark.scheduler.Task.run(Task.scala:109)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
ji-xin commented 5 years ago

natlog_log.txt openie_log.txt