Open piojanu opened 1 year ago
@piojanu what helps with the OOM issues with NVT is the part_size and the row group memory size of your parquet file(s). you can also repartition your dataset and save back to disk, and that might help with OOM. for LocalCUDACluster
args you can read here: https://docs.rapids.ai/api/dask-cuda/nightly/api/
if you have a single GPU you can try to set the row group size of your files and that would help without LocalCudaCluster.
There is a LocalCudaCluster example here: https://github.com/NVIDIA-Merlin/Merlin/blob/main/examples/quick_start/scripts/preproc/preprocessing.py
Hi!
I have follow up questions:
Thanks for help :)
By accident, I've found out that merlin.io.dataset.Dataset.shuffle_by_keys
is the root of this OOM.
ops.Categorify
OOM even when following the troubleshooting guide:
session_id
in BigQuery.
index
and calculate_divisions
to the dd.read_parquet
.nvt.ops.GroupBy
on the same data doesn't return the expected number of sessions.
Hi!
I've run into such a problem:
I run this code in JupyterLab on the GCP VM with NVIDIA V100 16GB GPU. I've also tried
nvtabular.utils.set_dask_client
and it didn't solve the problem.Questions: