all-in-one-docker-bigdataops is a comprehensive Docker Compose environment that simplifies Big Data operations by bundling Hadoop, Spark, Hive, Hue, and Airflow into a ready-to-run stack, with example workflows, quick setup, and easy customization, making it ideal for learning, development, and testing in Big DataOps.
Añadida linea en Dockerfile de spark para exportar la variable de entorno y así poder obtenerla en jupyter
This pull request includes updates to the Spark session Jupyter notebook to improve logging and execution metadata tracking, as well as a minor change to the Spark Dockerfile to set the SPARK_EXECUTOR_ID.
Añadida linea en Dockerfile de spark para exportar la variable de entorno y así poder obtenerla en jupyter This pull request includes updates to the Spark session Jupyter notebook to improve logging and execution metadata tracking, as well as a minor change to the Spark Dockerfile to set the
SPARK_EXECUTOR_ID
.Jupyter Notebook Updates:
jupyter/notebooks/spark/basics/01_spark_session.ipynb
, jupyter/notebooks/spark/basics/01_spark_session.ipynbL51-R66)SPARK_EXECUTOR_ID
. (jupyter/notebooks/spark/basics/01_spark_session.ipynb
, jupyter/notebooks/spark/basics/01_spark_session.ipynbL264-R337)Dockerfile Update:
SPARK_EXECUTOR_ID
to the hostname in the Spark environment configuration. (spark/Dockerfile
, spark/DockerfileR49-R51)