Closed tensor-tang closed 6 years ago
I meet this situation too. In addition, when i use the inference by multi-thread way with "export OPENBLAS_NUM_THREADS=1", the program may end with the "Aborted" error!
@ChinaLiuHao And as an addition, the "Abort" error is randomly encountered, not always appears.
The OCR CRNN_CTC service also has a large memory:
This should be the reason. Paddle would alloc max chunk size at the first time.
After debugging we can found there is a flag to choose how much memory we would like to use at the first time. Default it would use about 3.2%(1/32) of your total memory.
usage:
your_app --fraction_of_cpu_memory_to_use=0.1 # it would use 3.2% * 0.1 of total
The track back should be like this:
@ChinaLiuHao About the "Abort" issue, we can open another issue to discuss it. Thanks.
This is an issue of NLP online service.
When run inference, the memory usage is always kept as about 6G, which is definitely larger than actually needed.