Using a very small amount of data, s3.write_partitioned crashes DHC on OOM. The below script should work in 24G Heap. However, in the DHC Code Studio, you can hover over the heap status and see memory usage increase rapidly. Running with a _rowcount of 2000 will crash DHC with an OOM.
Decreasing the number of unique values for the partitioned key mitigates the problem. So the issue appears to be more about, unique partition key values, or combinations of multiple partition keys, than number of rows.
Using a very small amount of data,
s3.write_partitioned
crashes DHC on OOM. The below script should work in 24G Heap. However, in the DHC Code Studio, you can hover over the heap status and see memory usage increase rapidly. Running with a _rowcount of 2000 will crash DHC with an OOM.Decreasing the number of unique values for the partitioned key mitigates the problem. So the issue appears to be more about, unique partition key values, or combinations of multiple partition keys, than number of rows.