Open krisiasty opened 4 years ago
thanks for the report @krisiasty , unfortunately most of the artifacts in this repo are in a maintenance mode. i've labelled this as a bug, but i'm not sure how quickly someone would be able to address it.
PR #110 should handle this as we're adding Spark 3.0 with Hadoop 3.2 jars.
Currently used hadoop version (2.7.3) is way too old (released back in June 2017). One of the consequences is the missing support for "fs.s3a.path.style.access" property of s3a filesystem layer, which means the s3-compatible object store must be configured with virtual hosting for buckets. This in turn is not supported on the OpenShift Container Storage 4.4 (or at least it is not enabled by default and not properly documented how to configure this feature).
That means the Spark + Object Store example in the basic tutorial won't work on the latest OpenShift Container Platform (4.4) with OpenShift Container Storage.
Despite having the following settings:
the code from tutorial raises exception trying to connect to the bucket vi virtual host (http://bucket.s3endpoint/ instead of https://s3endpoint/bucket/):