Open pawelantczak opened 5 years ago
@pawelantczak did you ever figure this out?
and ... the solution is ?????? ... did you use docker ??
I encountered this problem while running spark on cluster mode. Based on my research and understanding, when spark is running on cluster mode, it tends to write on tmp file first which can be stored anywhere randomly on the worker nodes.
I also saw that for some people, the problem was solved by using the latest version, but it did not help me.
Running the spark on standalone mode solved this problem for me. To run spark in standalone mode just use this piece of code while creating the sparksession object:
SparkSession.builder()
.appName(yourAppName).master("local")
.getOrCreate()
Hope this helps.
Hello.
While in local mode everything is running smoothly, when I'm executing application on remote cluster, I'm getting this error:
in addition, when
tempLocation
is set, I can see files on Spark server.