Open cptera opened 2 years ago
@yuting-liu can you take a look as this is suspected to be related to the forked docker-java removal? Thanks
@cptera the error itself seems to be related to that the container didn't get restarted appropriately. Did you happen to see errors from the container log? It might include more details about the error.
Hmm, you're right, it looks like the container connector_healthcheck.1.r6cze5gx0o50en9a713u7h860
restarted, but the logs are still being collected after it restarted so I don't think it's an issue with docker or the container itself failing to restart.
Like we deploy everything in a docker swarm and we designed our containers to restart on certain errors and in this case it restarted on an error it should have. We've been running it this way for several years and we don't make major changes very often.
With the log4j bug our fix of "just using an older version" is no longer viable
We deploy the
sumologic/collector:latest
image in our customers production environemt to collect logs for our product (a set of containers running) and have it send logs to our sumologic account. Over the last few months we've noticed a lot of spam logs due to the sumologic collector throwing error likeFrom our investigation this is mostly happening for collector
19.351-4
, but it's also happening for19.361-4
and if I had to guess this may be due to switching from the Forked docker-java dependency to the open source one as listed in the release notes for https://github.com/SumoLogic/sumologic-collector-docker/releases/tag/v19.351-4Our main issue is that this causes a lot of our sumo quota to get used up, log collection seems to be functioning fine.