Open KevinyhZou opened 1 day ago
The plan:
CollectLimit 10
+- CHNativeColumnarToRow
+- ^(16) ProjectExecTransformer [2024-10-29 AS day#260, coalesce(name#273, all)#275 AS name#261]
+- ^(16) HashAggregateTransformer(keys=[2024-10-29#274, coalesce(name#273, all)#275], functions=[], isStreamingAgg=false)
+- ^(16) InputIteratorTransformer[2024-10-29#274, coalesce(name#273, all)#275]
+- ColumnarExchange hashpartitioning(2024-10-29#274, coalesce(name#273, all)#275, 1), ENSURE_REQUIREMENTS, [plan_id=1362], [shuffle_writer_type=hash], [OUTPUT] List(2024-10-29:StringType, coalesce(name#273, all):StringType)
+- ^(15) HashAggregateTransformer(keys=[2024-10-29#274, coalesce(name#273, all)#275], functions=[], isStreamingAgg=false)
+- ^(15) ExpandExecTransformer [[name#268, name#268, 2024-10-29#274, coalesce(name#273, all)#275], [null, name#268, 2024-10-29#274, coalesce(name#273, all)#275]], [name#273, name#268, 2024-10-29#274, coalesce(name#273, all)#275]
+- ^(15) !ProjectExecTransformer [name#268, 2024-10-29 AS 2024-10-29#274, coalesce(name#273, all) AS coalesce(name#273, all)#275]
+- ^(15) NativeFileScan parquet default.test_tbl3[name#268] Batched: true, DataFilters: [], Format: Parquet, Location: InMemoryFileIndex(1 paths)[hdfs://testcluster/user/hive/warehouse/test_tbl3], PartitionFilters: [], PushedFilters: [], ReadSchema: struct<name:string>
the plan seems not right on expand transformer, [name#268, name#268, 2024-10-29#274, coalesce(name#273, all)#275]
Check rule PushdownAggregatePreProjectionAheadExpand
Backend
CH (ClickHouse)
Bug description
test table schema: (id bigint, name string, day string)
query sql
exception message
Spark version
Spark-3.3.x
Spark configurations
No response
System information
No response
Relevant logs
No response