I am wondering if this is caused by my configuration or something else.
We are testing entity-fishing disambiguate model in Kubernetes 1.24.
Using grobid/entity-fishing:0.0.6 image for testing and we followed the instruction in here
e.g.
curl 'http://localhost:8090/service/disambiguate' -X POST -F "query={ 'text': 'The army, led by general Paul von Hindenburg defeated Russia in a series of battles collectively known as the First Battle of Tannenberg. But the failed Russian invasion, causing the fresh German troops to move to the east, allowed the tactical Allied victory at the First Battle of the Marne.', 'processSentence': [ 1 ], 'sentences': [ { 'offsetStart': 0, 'offsetEnd': 138 }, { 'offsetStart': 138, 'offsetEnd': 293 } ], 'entities': [ { 'rawName': 'Russian', 'type': 'NATIONAL', 'offsetStart': 153, 'offsetEnd': 160 } ] }"
the very first request to the service is talking close to 30 seconds. In this case below, it took 25 seconds.
Any subsequent requests after that take less than 100 ms.
We are hitting the service using readinessProbe to make sure the service is available, but after the pod is ready, the first request either from outside or inside of the docker image is taking a long time.
I am wondering if this is caused by my configuration or something else. We are testing entity-fishing disambiguate model in Kubernetes 1.24. Using grobid/entity-fishing:0.0.6 image for testing and we followed the instruction in here
e.g.
the very first request to the service is talking close to 30 seconds. In this case below, it took 25 seconds. Any subsequent requests after that take less than 100 ms. We are hitting the service using readinessProbe to make sure the service is available, but after the pod is ready, the first request either from outside or inside of the docker image is taking a long time.
the memory of the server is consistently growing as more requests are coming in. e.g. it is currently 21% but continue to go up and crash.
Any insight on these two issues would be appreciated!