Hello! I've found a performance issue in dataset.py: dataset = dataset.batch(batch_size)(here) should be called before dataset = dataset.map(_parse, num_parallel_calls=AUTOTUNE)(here), which would make your program more efficient.
Besides, you need to check the function _parse(here) called in dataset.map() whether to be affected or not to make the changed code work properly. For example, if _parse needs data with shape (x, y, z) as its input before fix, it would require data with shape (batch_size, x, y, z).
Looking forward to your reply. Btw, I am very glad to create a PR to fix it if you are too busy.
Hello! I've found a performance issue in dataset.py:
dataset = dataset.batch(batch_size)
(here) should be called beforedataset = dataset.map(_parse, num_parallel_calls=AUTOTUNE)
(here), which would make your program more efficient.Here is the tensorflow document to support it.
Besides, you need to check the function
_parse
(here) called indataset.map()
whether to be affected or not to make the changed code work properly. For example, if_parse
needs data with shape (x, y, z) as its input before fix, it would require data with shape (batch_size, x, y, z).Looking forward to your reply. Btw, I am very glad to create a PR to fix it if you are too busy.