Open jnguyen32 opened 5 years ago
Currently, the implementations of the TransactionEncoder and frequent itemset mining algorithms don't support chunking.
What may help though is using a sparse dataframe for frequent itemset and rule mining. For example, if you set .transform(X, sparse=True)
for the TransactionEncoder, it will return a sparse DataFrame.
It just occurs to me that sth like Dask dataframes, which have out-of-core support, could also work, but I have not tested this -- currently, we only use pandas DataFrames for testing
Per the example:
What's the best way to use chunking so that we can get a transformed dataframe that can be used for rule mining?