Open XYAskWhy opened 6 years ago
@XYAskWhy the error above is caused by the mistake on our part in the inference mode of unet
.
We are running always unet_padded
and unet_padded_tta
in the inference mode and didn't catch that typo. I would suggest that you run evaluate again with the unet_padded
on --chunk_size 5000
or unet_padded_tta
on smaller chunk size to fit it in memory when combining tta predictions. My advice is to go with --chunk_size 200
with unet_padded_tta
as it gives the best results.
@jakubczakon Many thanks, but training configuration might not be practical, since most mainstream GPUs now have about 10G memory while the 20 images batch only use 2G. As a result, training is very slow. What's your suggestion on larger batch_size and corresponding learning rate?
Very simple just change batch_size_train in the neptune.yaml. you change all other things there too. Including encoder network fron resnet34 to resnet152 or 101, learning rates training schedule and other stuff
You can also train multi gpu. Remember to set num_workers to a higher number because that usually is the bottleneck
how can you work Neptune in Offline Mode
Hi @hs0531
You can do something like this:
from neptune import OfflineBackend
neptune.init(backend=OfflineBackend())
...
as [explained here[(https://docs.neptune.ai/neptune-client/docs/neptune.html?highlight=offline).
In that case, nothing will be logged to Neptune -> I use it usually for debugging purposes.
thank you
---Original--- From: "Jakub"<notifications@github.com> Date: Wed, May 20, 2020 17:35 PM To: "neptune-ai/open-solution-mapping-challenge"<open-solution-mapping-challenge@noreply.github.com>; Cc: "hs0531"<348580064@qq.com>;"Mention"<mention@noreply.github.com>; Subject: Re: [neptune-ai/open-solution-mapping-challenge] predicting/evaluating issue (#123)
Hi @hs0531
You can do something like this: from neptune import OfflineBackend neptune.init(backend=OfflineBackend()) ...
as [explained here[(https://docs.neptune.ai/neptune-client/docs/neptune.html?highlight=offline).
In that case, nothing will be logged to Neptune -> I use it usually for debugging purposes.
— You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub, or unsubscribe.
thank you … ---Original--- From: "Jakub"<notifications@github.com> Date: Wed, May 20, 2020 17:35 PM To: "neptune-ai/open-solution-mapping-challenge"<open-solution-mapping-challenge@noreply.github.com>; Cc: "hs0531"<348580064@qq.com>;"Mention"<mention@noreply.github.com>; Subject: Re: [neptune-ai/open-solution-mapping-challenge] predicting/evaluating issue (#123) Hi @hs0531 You can do something like this: from neptune import OfflineBackend neptune.init(backend=OfflineBackend()) ... as [explained here[(https://docs.neptune.ai/neptune-client/docs/neptune.html?highlight=offline). In that case, nothing will be logged to Neptune -> I use it usually for debugging purposes. — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub, or unsubscribe.
which version of neptune do you install. i do follow you but get "cannot import name OfflineBackend"
Hi @hs0531
You can do something like this:
from neptune import OfflineBackend neptune.init(backend=OfflineBackend()) ...
as [explained here[(https://docs.neptune.ai/neptune-client/docs/neptune.html?highlight=offline).
In that case, nothing will be logged to Neptune -> I use it usually for debugging purposes.
which version of neptune do you install. i do follow you but get "cannot import name OfflineBackend"
When predicting or evaluating with
python main.py -- predict(evaluate) --pipeline_name unet --chunk_size 5000
, the following error occurs:The error above may be caused by using
--chunk_size 5000
, since the program crashes exactly after 5000 iterations(?). But even if I don't specify chunk_size and just runpython main.py -- predict --pipeline_name unet
, another error occurs, which is the same error when I simply runpython main.py -- train_evaluate_predict --pipeline_name unet --chunk_size 5000
as ReadMe suggests.