Open mokundong opened 5 years ago
You can refer to this. import predict_online function in run_classifier_predict_online.py. It will create a session and graph initializing at first, and you won't initial model every time you send a request
94
You can refer to this. import predict_online function in run_classifier_predict_online.py. It will create a session and graph initializing at first, and you won't initial model every time you send a request
thanks!
@mokundong You may refer to bert-as-service
for serving and deploying BERT models.
@mokundong You may refer to
bert-as-service
for serving and deploying BERT models.
it works for classifier task?
@mokundong You may refer to
bert-as-service
for serving and deploying BERT models.it works for classifier task?
It didn't work for me for Question answering
when use flask deploy a fine tuned- classification task online, each post will reload the model. this takes about 4 seconds to respond, is there any solutions to avoid reloading?