Open sylvaincom opened 1 month ago
Thank you for this interesting feedback.
For a warning, I am not sure what the right behavior would be.
So I suggest picking sane defaults:
put
and put_item
, offer control to limit dataframe size using the keyword argument dataframe_max_rows
. None
means no limit, default is 1000.numpy_array_max_length
wit default value 1000 as well).Then an enhancement that must examined in a wider context:
As suggested by @rouk1, we could think about using Perspective as a visualization tool unlocking many interesting use-cases, including the ability to store large amounts of data in a skore database..
Discussed with @fcharras that suggested to have a look at DeltaLake, and parquet files.
Issue
From the notebook, suppose that I load a large dataframe of shape (1M, 1k) let's say. I try to store it: it takes a lot of time.
Expectation
I would like a warning to the user that his file is too big, or maybe just store a subset of the big Dataframe (and warning that the display of the dashboard is truncated). Indeed, no one needs to visualize a dataframe of shape (1M, 1k) in the dashboard.