Open matthiasgomolka opened 9 months ago
Hi @matthiasgomolka thank you for letting us know, this is interesting to see. We'll put it into our backlog for review.
Had the same issue
GX Documentation:
row_condition='col("foo").notNull()' # foo is not null
Spark Documentation:
column.IsNotNull()
[see: https://spark.apache.org/docs/latest/api/python/reference/pyspark.sql/api/pyspark.sql.Column.isNotNull.html]
Traceback error:
raise gx_exceptions.MetricResolutionError(\ngreat_expectations.exceptions.exceptions.MetricResolutionError: unable to parse condition: col('column_name').notNULL()\n"
PLEEEEASE get rid of this row_condition='col("foo").notNull()'
and allow simple SQL syntax passthru: row_condition = 'fld1=5 OR fld2<>7 AND fld3 <9'
Describe the bug The documentation states that row conditions for Spark should be specified like this:
row_condition='col("foo") == "Two Two"'
.However, if I try it like this, I get this error:
Interestingly, it works like this:
To Reproduce great_expectations.yml:
Code:
Traceback:
Expected behavior I would expect that GX works as documented.
Environment (please complete the following information):