Closed siqueirarenan closed 2 years ago
I am having exactly the same issue. Is there any workaround for this problema as temp solution? I've tried some deleting of netty-* jars but nothing worked so far.
Same here as well. Wonder if it got updated or something else changed (like Docker version or something) because it was working a month ago and now it throws an error:
py4j.protocol.Py4JJavaError: An error occurred while calling None.org.apache.spark.api.java.JavaSparkContext.
: java.lang.NoSuchMethodError: io.netty.util.internal.ReflectionUtil.trySetAccessible(Ljava/lang/reflect/AccessibleObject;)Ljava/lang/Throwable;
Any help will be greatly appreciated
Update: I got it working by removing all netty
jars from both aws-glue-libs
and the spark jars
folder except netty-all-<>
jar.
@vstoyanoff Thanks for the comment. I can confirm that it works once you run below script in your notebook.
%%bash
find /home/spark-2.4.3-bin-spark-2.4.3-bin-hadoop2.8/jars/ -name "netty-*" ! -name 'netty-all*' -delete
find /home/aws-glue-libs/jarsv1/ -name "netty-*" ! -name 'netty-all*' -delete
Thank you for reporting this issue. We apologize for delay in response. Yes, it seems that the root cause is library conflict of netty jar files.
We resolved that issue in the Glue v2/v3 Docker images. Here's the blog post for that. https://aws.amazon.com/blogs/big-data/develop-and-test-aws-glue-version-3-0-jobs-locally-using-a-docker-container/
In case you still see the same issue in those newer images, please let us know. Thank you.
Dear all,
we are facing some problems when trying to run any spark command using the official AWS container for Glue V1. We are following this documentation https://aws.amazon.com/blogs/big-data/developing-aws-glue-etl-jobs-locally-using-a-container/.
When we simply try to run the command
spark-2.4.3-bin-spark-2.4.3-bin-hadoop2.8/bin/pyspark
to run the Spark CLI, we receive the following error.