Open king5holiday opened 1 year ago
Did you write the table using Flink, flink could create some imtermediate data files which starts with .
. And what verion of Hudi did you use then?
Flink
@danny0405 Thank you for your replay! yes, I wrote data to hudi by flink and the version of hudi is 0.11.1
Flink
@danny0405 Thank you for your replay! yes, I wrote data to hudi by flink and the version of hudi is 0.11.1
Can you use 0.12.3 or 0.13.1, it should be fixed.
Flink
@danny0405 Thank you for your replay! yes, I wrote data to hudi by flink and the version of hudi is 0.11.1
Can you use 0.12.3 or 0.13.1, it should be fixed.
@danny0405 thank you! Should hudi-spark-bundle version and hudi-flink-bundle version be changed to 0.12.3 or 0.13.1 at the same time?
Yeah, we better upgrade the bundle jars altogether.
Yeah, we better upgrade the bundle jars altogether.
I changed hudi-spark-bundle version from 0.11.1 to 0.13.1 and try to read data by spark, but the error reported again. when i use flink to read, although old version, it went well, so does flink has a filtering mechanism or remove the parquet files that start with dot(.)?
Did you read the table by specifying hudi as the format? Or just read it as a raw parquet table.
Did you read the table by specifying hudi as the format? Or just read it as a raw parquet table.
hi, there are some codes,
spark.read .format("org.apache.hudi") .load(HUDIPATH)
Not sure, maybe you can delete the hidden files manually, there is no atomatic fix when upgrading to 0.13.1.
Not sure, maybe you can delete the hidden files manually, there is no atomatic fix when upgrading to 0.13.1.
ok, maybe it's an effective way until now. Looking forward to fixing the bug in new version. Thank you!
so i want to know why there will be some parquet files which the file name start with dot(.) when I write data to hudi? And how to filter these files when I read hudi by spark? Thank you very much!