Open qingyuan18 opened 1 year ago
Did you ever try to write to a legacy table? It seems a version compatibility.
no , i have cleaned up table dataand re run the job . th error still reproduce
---Original--- From: "Danny @.> Date: Wed, Apr 5, 2023 13:48 PM To: @.>; Cc: @.**@.>; Subject: Re: [apache/hudi] [SUPPORT] hudi 0.12 spark batch ingestion throw outarchive format validation error (Issue #8382)
Did you ever try to write to a legacy table? It seems a version compatibility.
— Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you authored the thread.Message ID: @.***>
Did you also clean the .hoodie/archive folder ?
yes, indeed
---Original--- From: "Danny @.> Date: Wed, Apr 5, 2023 14:02 PM To: @.>; Cc: @.**@.>; Subject: Re: [apache/hudi] [SUPPORT] hudi 0.12 spark batch ingestion throw outarchive format validation error (Issue #8382)
Did you also clean the .hoodie/archive folder ?
— Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you authored the thread.Message ID: @.***>
It looks like a version compatibility issue, in old version Hudi, the archived entry does not have field: operationType
.
@qingyuan18 Were you able to resolve this issue? If yes can you share the resolution please.
Tips before filing an issue
Have you gone through our FAQs?
Join the mailing list to engage in conversations and get faster support at dev-subscribe@hudi.apache.org.
If you have triaged this as a bug, then file an issue directly.
Describe the problem you faced Hudi 0.12 ,spark 3.2, writeToHudiByNoPartition throw out exception:
seems like it doesn't recognize the hudi's archive metadata avro format
To Reproduce
Steps to reproduce the behavior:
read datasource with spark dataframe
config hudi write parameter as following
run the spark app with writeHudi function: writeToHudiByPartition( df2, sinkTable, sink_alliances_table_key, sink_alliances_distinct_field, "date_part", hiveDB, save_path)
after 30 commits, which trigger the archive process, it throw out the exception as ahead
Expected behavior
A clear and concise description of what you expected to happen.
Environment Description
Hudi version : 0.12.0
Spark version : 3.2.1
Hive version : 3.2.1
Hadoop version :
Storage (HDFS/S3/GCS..) : s3
Running on Docker? (yes/no) : no
Additional context
Add any other context about the problem here.
Stacktrace
Add the stacktrace of the error.