Closed conderls closed 9 months ago
Hi @conderls! I would prefer not to add such complexity to the library.
Considering that parquet-hadoop in version 1.10.1 is five years old, I recommend you upgrade it. You may not even need to upgrade Spark. Otherwise, you still have the option of shading.
Hi @conderls! I would prefer not to add such complexity to the library.
Considering that parquet-hadoop in version 1.10.1 is five years old, I recommend you upgrade it. You may not even need to upgrade Spark. Otherwise, you still have the option of shading.
getPath()
method breaks the compatibility and just return the path.parquet-hadoop
as provided scope, if without upgrade spark, then I may not able to upgrade parquet-hadoop
;parquet-hadoop
1.10.1 and parquet4s
depends on 1.13.1, the shading option may not work here?2) Check out https://hadoopsters.com/how-to-override-a-spark-dependency-in-client-or-cluster-mode-2860a64ad1d5 3) In this option, you should shade parquet4s' dep
as the parquet-hadoop
parquetVersion = "1.13.1"
and the spark3.1(spark-sql) with version1.10.1
, there some conflicts exists and lead to Errorthe
HadoopInputFile.getPath()
method is added since version1.12.x
, and it did nothing but return the path fromFileStatus
so, is it possible to make it more compatible by getting rid of calling
HadoopInputFile.getPath()
, and usingnew Path(HadoopInputFile.toString)
instead?