Closed youngDouble closed 5 years ago
Hi @youngDouble,
so the log output is hanging at "Epoch 77"? Since you are running with nohup, are you sure that the process is still running? It could have crashed, for example on unsufficient memory. Can you try running the training again?
I would also suggest running with less epochs, e.g. 100, since the accuracy seems quite high already on epoch 77. Also, if you want to try reserving part of training data for validation, you can set "validation_size" in your JSON file. For example, you can use "validation_size": 0.2 for validating your model on 20% of your data. Then, both training and validation metrics will be shown during training.
Also, please see the updated README section on training: https://github.com/Merck/deepbgc#train-deepbgc-on-your-own-data
You can upgrade DeepBGC to version 0.1.10 to be able to provide your trained model for detection and classification like so:
deepbgc pipeline \
mySequence.fa \
--detector path/to/myDetector.pkl \
--classifier path/to/myClassifier.pkl
Thank you for your reply, as you said, it is a problem with insufficient memory. After adjusting the parameters(validation_size
and num_epochs
), it can work.
It seems that the training process uses a lot of memory (more than 256G), I have delivered the task to the cluster.
Hello! Recently I used
deepbgc
in my work, but I encountered some problems while training my model. My code:It can start running, but after a while, he seems to be deadlocked: The end of the log :
This state has been going on for a long time (about 2 days, my input file is only 2M, including 515 BGC) What can I do? thank you!