heirinsinho / all-in-one-docker-bigdataops

all-in-one-docker-bigdataops is a comprehensive Docker Compose environment that simplifies Big Data operations by bundling Hadoop, Spark, Hive, Hue, and Airflow into a ready-to-run stack, with example workflows, quick setup, and easy customization, making it ideal for learning, development, and testing in Big DataOps.
10 stars 3 forks source link

Funcion get_executor_partition_info() arreglada #3

Closed JuanCaarPapers closed 1 month ago

JuanCaarPapers commented 1 month ago

Añadida linea en Dockerfile de spark para exportar la variable de entorno y así poder obtenerla en jupyter This pull request includes updates to the Spark session Jupyter notebook to improve logging and execution metadata tracking, as well as a minor change to the Spark Dockerfile to set the SPARK_EXECUTOR_ID.

Jupyter Notebook Updates:

Dockerfile Update:

heirinsinho commented 1 month ago

No sirve. Esto identifica el worker, que ya lo hicimos en clase, pero no al ejecutor