An error that is likely to be caused by the dependency parser.
2018-11-12 16:46:10 INFO MaxentTagger:88 - Loading POS tagger from edu/stanford/nlp/models/pos-tagger/english-left3words/english-left3words-distsim.tagger ... done [0.5 sec].
2018-11-12 16:46:10 INFO DependencyParser:88 - Loading depparse model: edu/stanford/nlp/models/parser/nndep/english_UD.gz ...
2018-11-12 16:46:16 ERROR Utils:91 - Aborting task
java.lang.OutOfMemoryError: Java heap space
at edu.stanford.nlp.parser.nndep.Classifier.preCompute(Classifier.java:662)
at edu.stanford.nlp.parser.nndep.Classifier.preCompute(Classifier.java:644)
at edu.stanford.nlp.parser.nndep.DependencyParser.initialize(DependencyParser.java:1189)
at edu.stanford.nlp.parser.nndep.DependencyParser.loadModelFile(DependencyParser.java:630)
at edu.stanford.nlp.parser.nndep.DependencyParser.loadFromModelFile(DependencyParser.java:499)
at edu.stanford.nlp.pipeline.DependencyParseAnnotator.<init>(DependencyParseAnnotator.java:57)
at edu.stanford.nlp.pipeline.AnnotatorImplementations.dependencies(AnnotatorImplementations.java:240)
at edu.stanford.nlp.simple.Document$5.lambda$null$0(Document.java:147)
at edu.stanford.nlp.simple.Document$5$$Lambda$63/1641616399.get(Unknown Source)
at edu.stanford.nlp.util.Lazy$2.compute(Lazy.java:106)
at edu.stanford.nlp.util.Lazy.get(Lazy.java:31)
at edu.stanford.nlp.simple.Document$5.get(Document.java:147)
at edu.stanford.nlp.simple.Document$5.get(Document.java:142)
at edu.stanford.nlp.simple.Document.runDepparse(Document.java:918)
at edu.stanford.nlp.simple.Document.runNatlog(Document.java:938)
at edu.stanford.nlp.simple.Sentence.natlogPolarities(Sentence.java:897)
at edu.stanford.nlp.simple.Sentence.natlogPolarities(Sentence.java:905)
at ca.uwaterloo.cs651.project.SimpleNLP.lambda$main$71c840f9$1(SimpleNLP.java:114)
at ca.uwaterloo.cs651.project.SimpleNLP$$Lambda$27/273072814.call(Unknown Source)
at org.apache.spark.api.java.JavaPairRDD$$anonfun$toScalaFunction$1.apply(JavaPairRDD.scala:1040)
at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
at org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$4.apply(SparkHadoopWriter.scala:125)
at org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$4.apply(SparkHadoopWriter.scala:123)
at org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1414)
at org.apache.spark.internal.io.SparkHadoopWriter$.org$apache$spark$internal$io$SparkHadoopWriter$$executeTask(SparkHadoopWriter.scala:135)
at org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$3.apply(SparkHadoopWriter.scala:79)
at org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$3.apply(SparkHadoopWriter.scala:78)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
at org.apache.spark.scheduler.Task.run(Task.scala:109)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
An error that is likely to be caused by the dependency parser.