Closed vincenzon closed 1 hour ago
It looks like this is the problem:
I think it should be:
unquoted_identifier = Word(alphas + "_", alphanums + "_$")
thanks for reporting this. do you know if a leading underscore in column name is valid in spark sql?
According to this: https://spark.apache.org/docs/latest/sql-ref-identifier.html it is allowed. In fact, the way quoting is handled by pyiceberg is wrong on two levels:
Fixing the quote character is easy, fixing the second issue would be more involved.
Apache Iceberg version
0.8.0 (latest release)
Please describe the bug 🐞
A row_filter passed table overwrite throws a parse exception if the column begins with a underscore. The example below demonstrates the issue. I tried quoting the column name but that didn't help.