apache-spark-on-k8s / spark

Apache Spark enhanced with native Kubernetes scheduler back-end: NOTE this repository is being ARCHIVED as all new development for the kubernetes scheduler back-end is now on https://github.com/apache/spark/
https://spark.apache.org/
Apache License 2.0
612 stars 118 forks source link

tutorial spark-pi invocation failing in openshift #537

Closed erikerlandson closed 6 years ago

erikerlandson commented 6 years ago

Not sure what's going wrong - it seems to be complaining about not finding jar file in executor, although it's using the local://.... spark-pi example jars as the dep.

This log fragment from the driver looks suspicious:

2017-10-31 19:56:29 INFO  SparkContext:54 - Added JAR /opt/spark/examples/jars/spark-examples_2.11-2.2.0-k8s-0.5.0.jar at spark://spark-pi-1509479739170-driver-svc.webuisecure.svc.cluster.local:7078/jars/spark-examples_2.11-2.2.0-k8s-0.5.0.jar with timestamp 1509479789779
2017-10-31 19:56:29 WARN  KubernetesClusterManager:66 - The executor's init-container config map was not specified. Executors will therefore not attempt to fetch remote or submitted dependencies.
2017-10-31 19:56:29 WARN  KubernetesClusterManager:66 - The executor's init-container config map key was not specified. Executors will therefore not attempt to fetch remote or submitted dependencies.
erikerlandson commented 6 years ago

Problem appears to be that the process of copying the jar from driver to executors was failing due to the container not running as root. A typical mismatch between default kube behavior and default openshift behavior.