Open indygriffiths opened 1 year ago
For the time being we've worked around this by splitting our one Fleet Server host with two URLs in KIbana into two separate ones for each URL, and assigning this machine to a different agent policy with the new Fleet host entry. Not the prettiest solution but seems to work.
I can see traffic in our proxy logs that the request is making it to the proxy so I'm not sure why the second URL is failing.
This is odd, it suggests the agent proxy configuration is working correctly. It is possible the timeout is applying to both requests sequentially instead of each one individually, but I think the logic here is the same on 8.5.3 as it is on 8.6.2.
Isn't it fixed by https://github.com/elastic/elastic-agent/pull/2239 ?
We have two Fleet Server URLs configured:
https://<private>-logs.fleet.vpce.ap-southeast-2.aws.elastic-cloud.com:443
)https://<private>-logs.fleet.ap-southeast-2.aws.found.io:443
)On a Ubuntu server that uses a proxy server to pass requests to Elastic (configured through envvars set by systemd), attempting to upgrade the agent from 8.5.3 to 8.6.2 fails during the step to enroll the new agent to Fleet Server. The first URL fails as expected for this server, but the second endpoint which is accessible on the server through the proxy fails with a
context canceled
error, which causes the agent to roll back to the old version.I can see traffic in our proxy logs that the request is making it to the proxy so I'm not sure why the second URL is failing. The request duration in the logs also seem off since it seems like it's timing out after a minute.
Since it's Fleet-managed I also haven't figured out how to increase the log verbosity and have it apply to the upgraded version.