Closed warrenronsiek closed 7 years ago
I am facing the exact same issue. Can someone please help out here?
@warrenronsiek were you able to resolve this issue?
@saurabh14rajput I ended up not using this library. Instead I created workaround with a UDF that wrapped the features of Stanford nlp that I wanted to use. Probably not super efficient or best practice - but it turns out to be relatively fast.
Okay. Thanks!
Hey, why do you include the corenlp in you dependencies? It is already done by spark-corenlp, see https://github.com/databricks/spark-corenlp/blob/master/build.sbt#L39
. Can you try to use version corenlp 3.6.0 and report back if you still have issues?
Hey, I think the issue is that you are mixing spark-corenlp 2.11
with Spark 2.10
. Your should replace the line
libraryDependencies += "databricks" % "spark-corenlp" % "0.2.0-s_2.11"
with
libraryDependencies += "databricks" % "spark-corenlp" % "0.2.0-s_2.10"
I'm getting the same error. I cloned this repo, ran sbt package
to build the jar, then invoked spark shell like this:
/opt/spark/spark-2.0.1/bin/spark-shell --jars ~/spark-corenlp_2.10-0.3.0-SNAPSHOT.jar
I get the error even if I specify library dependencies, like this:
/opt/spark/spark-2.0.1/bin/spark-shell --jars ~/spark-corenlp_2.10-0.3.0-SNAPSHOT.jar --packages databricks:spark-corenlp:0.2.0-s_2.10,edu.stanford.nlp:stanford-corenlp:3.7.0
@iandow try with spark-corenlp_2.11
. Spark 2.x is using Scala 2.11.
No dice. Same error.
@zouzias using
libraryDependencies += "databricks" % "spark-corenlp" % "0.2.0-s_2.10"
solved the problem for the example I posted above. I cant speak for the other people who are getting the same error.
I'm getting the same error that raghugvt posted here. He solved the problem by bundeling everything together in one jar, however thats not an option as I would like to use spark-corenlp in a notebook.
My build.sbt is as follows:
I'm testing with this script:
Which results in the error:
What is going wrong here?