Open jevy-wangfei opened 6 years ago
Thanks for the report @jevy-wangfei - are you able to share the details of what it was trying to index when it did this? This might be related to https://github.com/magda-io/magda/issues/1068
As specified in issue #1068 , we modified the MAX_EVENTS
to 10
instead of 100
, and was trying re-index all of the datasets harvested from about 30 data sources ( ~70K datasets). From the monitor we found that the indexer would consume really a lot memory continuously without release it. Because of the resource limitation of K8S, indexer will be stoped by K8S when it consume 4.6Gb memory.
(Enlarge VM memory usage by adding an environment param JAVA_OPTS in indexer helm chart:
env:
- name: JAVA_OPTS
value: -Xmx6114M -Xms6114M -XX:+CMSClassUnloadingEnabled -XX:MaxGCPauseMillis=1000
-XX:+UseG1GC -XX:GCTimeRatio=3
)
After switching the indexer to your team pre-builded data61/indexer/v0.0.41-0 (limit DAP connector to harvest only 24 distributions), the indexer could index all of our data, but it consumed 4.4GB memory and never releassed.
So, this may be a bug.
Problem description
Indexer consume really huge memory when re-indexing data, and the indexer pod will exist when it reaches the pod limitation. But when set the pod without limitation on memory and CPU usage, it will still exist with OutOfMemory Error
Problem reproduction steps
Log of existing when reach limitation
Log of existing with OutOfMemory Error
Screenshot / Design / File reference
Screenshot of existing when reach limitation
Screenshot of existing with OutOfMemory Error