Closed bbetter173 closed 5 years ago
If the purpose of the check process is to check external availability of the acme challenge, perhaps making the request to the external ingress IP would be the best idea?
I'm unsure what the bug is here - according the ACME spec, cert-manager should always communicate with a challenge endpoint using HTTP. HTTPS is specifically not supported and should not be used here.
cert-manager already attempts to make a request to the external address for the ingress controller (i.e. by making a request to the same IP that Letsencrypt will use when performing a challenge validation).
Can you describe your DNS/ingress further? It sounds like you have some kind of split horizon DNS set up that may be tripping up cert-manager?
The challenge is being requested over HTTP, but the nginx-ingress controller is expecting requests to be made using the proxy protocol - which my load balancer is configured to do.
When the go-http client makes a request directly to the nginx-ingress controller (i.e. not using the load balancers external IP) the proxy_protocol isn't used, causing it to fail,
If I manually run curl -v 'http://therealhost.example.com/.well-known/acme-challenge/9oQ5DbRUHNpnIsqvlvFUcb-km2OgpckyaXXEQh9cQQk
the response works fine, as the initial HTTP request is terminated by the load balancer, and the backend request to nginx-ingress is made using the proxy_protocol.
I am currently using the quay.io/kubernetes-ingress-controller/nginx-ingress-controller:0.12.0
controller, but this is not related to HTTPS, or redirection in any manner.
@munnerz - There is no split horizon DNS, the only issue is that the nginx-ingress controller is expecting requests using the proxy_protocol method, and the cert-manager controller is making requests to it using plain HTTP.
I would suggest adding a flag that toggles cert-manager to make all requests to the nginx-ingress controller via the load balancer instead of hitting it directly.
More troubleshooting indicates there is some strange behaviour in my cluster, using a similar setup in minikube works fine.
The error message I'm seeing when I set logging to verbose is: I0414 05:41:23.649175 1 http.go:410] ACME HTTP01 self check failed for domain "therealhost.example.com", waiting 5s: Get http://therealhost.example.com/.well-known/acme-challenge/9oQ5DbRUHNpnIsqvlvFUcb-km2OgpckyaXXEQh9cQQk: EOF
From within the cert-manager instance, the DNS host resolves correctly:
kubectl exec -it --namespace=shared-services cert-manager-cert-manager-5c7bfd7dc4-kpffn sh
/ # nslookup therealhost.example.com
nslookup: can't resolve '(null)': Name does not resolve
Name: therealhost.example.com
Address 1: 123.123.123.123
But if I install curl within the cert-manager instance I get odd behaviour:
/ # curl 'http://therealhost.example.com/.well-known/acme-challenge/9oQ5DbRUHNpnIsqvlvFUcb-km2OgpckyaXXEQh9cQQk' -v
* Trying 103.75.202.143...
* TCP_NODELAY set
* Connected to therealhost.example.com (123.123.123.123) port 80 (#0)
> GET /.well-known/acme-challenge/9oQ5DbRUHNpnIsqvlvFUcb-km2OgpckyaXXEQh9cQQk HTTP/1.1
> Host: therealhost.example.com
> User-Agent: curl/7.59.0
> Accept: */*
>
* Empty reply from server
* Connection #0 to host therealhost.example.com left intact
curl: (52) Empty reply from server
Whereas if I run the same command from any other host I get this:
curl 'http://therealhost.example.com/.well-known/acme-challenge/9oQ5DbRUHNpnIsqvlvFUcb-km2OgpckyaXXEQh9cQQk' -v
* About to connect() to therealhost.example.com port 80 (#0)
* Trying 103.75.202.143...
* Connected to therealhost.example.com (123.123.123.123) port 80 (#0)
> GET /.well-known/acme-challenge/9oQ5DbRUHNpnIsqvlvFUcb-km2OgpckyaXXEQh9cQQk HTTP/1.1
> User-Agent: curl/7.29.0
> Host: therealhost.example.com
> Accept: */*
>
< HTTP/1.1 200 OK
< Server: nginx/1.13.9
< Date: Sat, 14 Apr 2018 05:49:08 GMT
< Content-Type: text/plain; charset=utf-8
< Content-Length: 87
< Connection: keep-alive
<
* Connection #0 to host therealhost.example.com left intact
9oQ5DbRUHNpnIsqvlvFUcb-km2OgpckyaXXEQh9cQQk.zrZMCbhC-Lh8qDFGdhEMA4BvJkeBPAzkThQKl4U_sOE
@HeWhoWas i ran into the same issue today, did you find a solution?
from within cert-manager pod, DNS lookups resolve the public loadbalancer IP.
but it still seems that cert-manager will contact the node-ip directly to solve the acme challenge, hence does not go through the public loadbalancer
@metallhopf Our solution was to use DNS-01 validation, bypassing these issues altogether. I believe at the time this ticket was opened, that feature was only just being introduced and wasn't an option in a release version.
Sorry I can't be more help, but for what is worth we've found DNS validation to be simpler and more robust than HTTP.
I would suggest adding a flag that toggles cert-manager to make all requests to the nginx-ingress controller via the load balancer instead of hitting it directly.
That what CM always tries to do. It hits the domain name of the certificate using HTTP. It expects that domain name will resolve to the external IP. And that's the point, to test the challenge as e.g. Let's Encrypt would, from the outside. There is no value or point in CM making extra steps to find out and use the internal node IP instead, that wouldn't be checking anything useful.
If the traffic from the CM or CM challenge Pods are not going external, probably you have some special cluster DNS, network DNS, or perhaps hairpin routing getting in the way.
If the traffic from the CM or CM challenge Pods are not going external, probably you have some special cluster DNS, network DNS, or perhaps hairpin routing getting in the way.
you are correct, i was able to confirm this. the public loadbalancer ip is resolved correctly for the http request.
but traffic is routed internally directly from the cert-manager pod to nginx-ingress-controller (does not go through the loadbalancer).
cluster is hosted on digitalocean, will edit this post if i find a solution.
If the traffic from the CM or CM challenge Pods are not going external, probably you have some special cluster DNS, network DNS, or perhaps hairpin routing getting in the way.
you are correct, i was able to confirm this. the public loadbalancer ip is resolved correctly for the http request.
but traffic is routed internally directly from the cert-manager pod to nginx-ingress-controller (does not go through the loadbalancer).
cluster is hosted on digitalocean, will edit this post if i find a solution.
got the same issue with proxy protocol at digitalocean. my only workaround is to temporarily disable proxy protocol on the load balancer (and nginx ingress config map) allowing the certificate to be issued.
i'm hoping for a better solution to avoid interruption in service during certificate renewal.
I get the same on a cluster at Brightbox, which also uses hairpinning by default and an external load balancer that sends PROXY protocol. The hairpinning causes other issues too admittedly, but it's obviously a common setting for Kubernetes clusters.
If the traffic from the CM or CM challenge Pods are not going external, probably you have some special cluster DNS, network DNS, or perhaps hairpin routing getting in the way.
Since it's fairly common (or at least, so I believe), perhaps an option just to disable testing the challenge? The challenge will work externally, from Let's Encrypt, just not internally.
Just hit this with a DO managed Kube cluster and a DO LB in proxy-protocol mode. Seems like the the nginx-ingress LB is broken with DO's proxy protocol:
https://github.com/kubernetes/ingress-nginx/issues/3996
From the Ingress
" while reading PROXY protocol, client: 10.244.0.1, server: 0.0.0.0:80
2019/04/13 15:28:02 [error] 1384#1384: *248667 broken header: "GET /.well-known/acme-challenge/oF_X6SITHseBK1hdpEiZKbVCkmjvIiTHlPO46XXsSJM HTTP/1.1
Host: example.com
User-Agent: Go-http-client/1.1
Accept-Encoding: gzip
Connection: close
Unfortunately the DNS01 challenge is broken for DigitalOcean in 0.7.0 (and based on my testing in 0.6.0 as well) so HTTP01 is a must for DO.
I faced this one as well, why is this one closed? Cert manager http01 challenge is not working with DO load balancer with proxy protocol at the moment, which is the only load balancer that makes sense (as it forwards request IP). I guess I will try to use DNS01 until resolved, but how HTTP01 should be working, can we open this one?
I faced this one as well, why is this one closed? Cert manager http01 challenge is not working with DO load balancer with proxy protocol at the moment, which is the only load balancer that makes sense (as it forwards request IP). I guess I will try to use DNS01 until resolved, but how HTTP01 should be working, can we open this one?
Unfortunately, DigitalOcean doesn't support DNS01. Maybe unless you're using DigitalOcean's DNS service - TBD.
@altoning Not sure what are trying to say, I managed to get digitalocean dns01 challenge working with cert manager 0.7.0 and proxy protocol load balancer. That being said, this one still needs to get attention, as this is just a temp solution, I don't wanna have digital ocean API key in k8s, it is tied to specific digital ocean user and it stops working when the user is removed from the project (that is by design and should be so, I just don't want to tie cert manager validation to that).
@HeWhoWas Please can you reopen this issue as it hasn't been resolved.
I believe this is due to a design flaw in Kubernetes. When using LoadBalancer services that use an external service (like Brightbox or DO mentioned here), kube-proxy intercepts the outgoing requests to the load balancer external IP at the network level to keep them within the Kubernetes cluster but doesn't understand that some LoadBalancers can do more than just standard TCP balancing. So this will break internal connections to external load balancers that do more, such as proxy-support or even SSL offloading.
We've now fixed this at Brightbox by not telling kube-proxy about the external IP addresses of the LoadBalancers, so it doesn't intercept them. I think DO are going to fix it the same way, and AWS have done this all along. See https://github.com/kubernetes/kubernetes/issues/66607 for more details.
So this isn't a cert-manager problem.
Reopening due to all the conversation and at @dottodot request. If this turns out to not be a cert manager issue, happy to have it closed.
Not sure this will help anyone, but I managed to fix this :) I removed the proxy-protocol from the nginx configmap. You can still put it on individual ingress resources though.
I think @johnl's reference to https://github.com/kubernetes/kubernetes/issues/66607 is the cause. By luck or design, it works on AWS because the AWS k8s cloud provider code only adds the external host name not the external IP address.
The good news is a patch to Kubernetes in https://github.com/kubernetes/kubernetes/pull/77523 should eventually fix this for everyone.
I don't think it is a cert-manager
problem, but it worth keeping this open to track the fix and warn other cert-manager
users of this issue.
@sorenmat is it possible you could share how you configure it on a per ingress basis?
sure @Philio
annotations:
kubernetes.io/ingress.class: "nginx"
use-proxy-protocol: "true"
@sorenmat Thanks for the quick reply, unfortunately though it didn't work, no proxy protocol configuration at all was added to nginx.conf
. Looking at the template it is using only global configuration (although I may have interpreted it wrong) for proxy protocol. Perhaps it's version specific, which version did you get it to work with?
Same issue here. I need proxy-protocol because of client IPs so it is not a solution to disable it. If it is on, cert-manager is not working because of that pre-check.
There are two solutions:
I would prefer solution No.1 because there is a reason we are checking that endpoint before asking LetsEncrypt to do the same. This check is valuable as it prevents quota issues.
This is important to be resolved as people will need proxy protocol and it is bad if cert-manager
is unable to work in that case. I think any of two solutions would let us go around this problem.
I see that pre-check code is at: https://github.com/jetstack/cert-manager/blob/70bc3e845bffac5acc10934911648a42a3a05ed1/pkg/issuer/acme/http/http.go#L184
With curl
, I was able to connect with this line. It returned body.
http_proxy=http://my.website.com:80 curl -v http://my.website.com/.well-known/acme-challenge/mwbIQcwaB9LL6wwIGRjQuRfL8cl5lFfGXocuQ3Y_fqs --haproxy-protocol
I see a recent commit https://github.com/jetstack/cert-manager/commit/099abed3fc6717010e2ed5bac795905c0bc0ebd0 and PR https://github.com/jetstack/cert-manager/pull/1850 by @kinolaev that can have connection to what we are talking about.
Is this change adding PROXY header to checks or just using proxy?
Hello @MichaelOrtho, yes, my PR can help by allowing to send self-check request to acme-solver service to bypass ingress controller. But in this case it’s looks like temporary solution, actual problem (as already mentioned) discussed here https://github.com/kubernetes/kubernetes/pull/77523: we need send traffic that targeting balancer but not yet passed through balancer to balancer. Difficult is to differ traffic from balancer and traffic from pods and nodes because we have many different types of balancers, for example, simple bgp/arp-based (like metallb)and feature-reach that can proxy-protocol and ssl-termination.
Hello @MichaelOrtho, yes, my PR can help by allowing to send self-check request to acme-solver service to bypass ingress controller. But in this case it’s looks like temporary solution, actual problem (as already mentioned) discussed here kubernetes/kubernetes#77523: we need send traffic that targeting balancer but not yet passed through balancer to balancer. Difficult is to differ traffic from balancer and traffic from pods and nodes because we have many different types of balancers, for example, simple bgp/arp-based (like metallb)and feature-reach that can proxy-protocol and ssl-termination.
Hi @kinolaev, I think that provider mentioned in this issue was Digital Ocean and I think they use AWS infrastructure so those using AWS have the same problem. I understand that it is only one of types of load balancer. But, we need to solve it or it will impact usage of cert-manager.
By the way... I see that your linked issue in Kubernetes became a part of v1.15.1.
Kubernetes ticket: https://github.com/kubernetes/kubernetes/pull/77523 Kubernetes changelog: https://github.com/kubernetes/kubernetes/blob/master/CHANGELOG-1.15.md
@MichaelOrtho people using vanilla AWS support never experienced this problem, because the AWS in-tree support happened to be using the workaround, as I explained above https://github.com/jetstack/cert-manager/issues/466#issuecomment-490275120
Digital Ocean did have the problem I think, because they were doing their own thing (even if on AWS), which triggered the problem. However I believe they already made their own patch to work around it the same way.
v1.15.1 should fix it for everyone and the workaround won’t be required.
I contacted DigitalOcean support and they've told me that it is expected to have Kubernetes v15.1 available in DO by early August. I've referenced this issue and also issue in Kubernetes so they are aware of importance of this release.
Hi there!
Thank you for contacting DigitalOcean and providing such a detailed ticket!
Our plans are to have 1.15.1 available by early August.
Let me know if you have any additional questions.
Regards,
John K. (redacted for security)
Senior Developer Support Engineer
Seems like it's been delayed for some reason.
Thank you for contacting DigitalOcean! We do plan on offering v1.15 to our customers before the end of August.
I created a fresh cluster on digitalocean using the latest kubernetes 1.15.2 and I am still encountering this issue. Can anyone confirm that it is working for them on DO managed kubernetes v1.15.2?
I created a fresh cluster on digitalocean using the latest kubernetes 1.15.2 and I am still encountering this issue. Can anyone confirm that it is working for them on DO managed kubernetes v1.15.2?
no. did the same few hours ago with negative result :(
Hi, Timo here from DigitalOcean. Unfortunately, kubernetes/kubernetes#77523 does not seem to fix kubernetes/kubernetes#66607. (See also my coworker's comment at https://github.com/kubernetes/kubernetes/pull/77523#issuecomment-490329492 and Andrew Sy Kim's follow up at https://github.com/kubernetes/kubernetes/pull/77523#issuecomment-490592181 on that PR.) AFAICT, another upstream change would be required to fix the issue.
We do offer a workaround in the latest releases of DOKS clusters. See my comment here for further context and documentation.
@timoreimann thanks for your comment. please explain - can I set annotation service.beta.kubernetes.io/do-loadbalancer-hostname
multiple times for single LoadBalancer
? (or comma separated?)
@serafim no you cannot, but you can create extra DNS records on your own all pointing at the same load balancer IP address.
I recently updated the corresponding CCM documentation section. Please have a look and let me know if that clarifies things.
@timoreimann thank you! This solved my issue with cert-manager when multiple domains pointed to one DNS record (load balancer's IP).
Hi there - we're going through the issue backlog and this has been around for a while. Based on my understanding of this issue, there's not any action in cert-manager to be taken.
It seems that the loadbalancer in front of your Ingress controller should be responsible for adding the PROXY header.
When cert-manager performs the 'self check', it does so by attempting to access your domain name just like any other user - i.e. it does not route traffic directly to your ingress controller and bypass the load balancer.
Unless there's some funky network stuff going on here, or a DNS server rewriting responses, this should just work as your load balancer will add the appropriate PROXY headers when talking to your ingress controller.
I'm going to close this now as I don't think there's anything actionable. If my understanding above is incorrect, please let me know 😄 but I don't see any reason why cert-manager should have to add these headers, as Let's Encrypt themselves will never in any case add these headers.
But the code to achieve proxy protocol is super tiny. Can't cert-manager just accept a configuration option so this would be opt-in?
Bumping this again as I have some time over the next few weeks to create a PR for this. Would like a confirmation that it would not be useless work before starting though.
@munnerz Would cert-manager accept the changes adding proxy-protocol mode for self-check behind a configuration parameter so the behaviour would be opt in?
@Jyrno42 Just tested this with a site behind nginx-ingress in proxy mode. Used images: jyrno42/cert-manager-controller-amd64:canary jyrno42/cert-manager-webhook-amd64:canary jyrno42/cert-manager-cainjector-amd64:canary
Self-check succeeded and the cert was assigned 🚀 Definitely not useless as the only way I can interact with an OpenStack Octavia load balancer properly is via proxy mode.
The purpose of the self check is to test an external request, to ensure when the ACME server makes its external request it will work. If your self-check requests are not going external, then I humbly submit that that is the problem that needs fixing.
@sfxworks You are my hero :) Your images works perfectly!
Not my images. Just fished them out from @Jyrno42 's repo. But hey yw for the convenience haha
I configured Gitlab to make automated releases with my patch every time a new version of cert-manager gets released.
Repository itself is here: https://gitlab.com/jyrno42/cert-manager-patcher
This means you can update your deployments to use v0.12.0 tag not canary as you might have used previously.
@Jyrno42 The posted gitlab link is a 404! Could you please check if the project is not private?
@Jyrno42, accessing repo https://gitlab.com/jyrno42/cert-manager-patcher, returns 404.
Sorry about that, now should be fixed! @Isakgicu @tareksamni
Was transferring it to my personal gitlab account instead of the company one and forgot to make it public.
Thank you @Jyrno42, it works perfectly!
Good job @Jyrno42 :) But I would really like to have this fix/option in the original cert-manager release.
Like many others here, we have to temporary deactivate proxy-protocol while we wait for the certificates to renew. Doing this in a production environment is not ideal.
We cant use dns01 since we do not own many of the domains used.
It looks like there is a KEP and corresponding PR to fix the underlying issue causing this in Kubernetes: https://github.com/kubernetes/enhancements/pull/1392
I'm still not convinced we should send PROXY headers during the self check or look to support this natively in cert-manager. Ultimately it is caused by traffic being routed incorrectly within your network/cluster, and once this issue is resolved there will be no good reason for something like this. If we were to add this, it'd need to be behind a feature gate, marked 'alpha', clearly link to the issues describing the problem and additionally noting that the feature will be removed in a future release 🙂 (and a note/FAQ in the docs about it and when you'd want to use it).
On Thu, 20 Feb 2020 at 12:50, Daniel Bjørnådal notifications@github.com wrote:
Good job @Jyrno42 https://github.com/Jyrno42 :) But I would really like to have this fix/option in the original cert-manager release.
Like many others here, we have to temporary deactivate proxy-protocol while we wait for the certificates to renew. Doing this in a production environment is not ideal.
We cant use dns01 since we do not own many of the domains used.
— You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/jetstack/cert-manager/issues/466?email_source=notifications&email_token=AABRWP6AJVBZLW2PBCLO6ZDRDZ4AXA5CNFSM4E2NPCQKYY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOEMNYEPY#issuecomment-589005375, or unsubscribe https://github.com/notifications/unsubscribe-auth/AABRWP4RIDNEEKSF4H67MQLRDZ4AXANCNFSM4E2NPCQA .
Is this a BUG REPORT or FEATURE REQUEST?:
/kind bug
What happened:
When running ingress-nginx with
use-proxy-protocol: true
, the check stage of cert-manager fails as it (appears to) communicate with the ingress controller using plain HTTP requests.What you expected to happen:
How to reproduce it (as minimally and precisely as possible):
use-proxy-protocol: true
, andproxy-real-ip-cidr: x.x.x.x
(Use the real load balancer IP) for the nginx controllerAnything else we need to know?:
Environment:
Client Version: version.Info{Major:"1", Minor:"10", GitVersion:"v1.10.0", GitCommit:"fc32d2f3698e36b93322a3465f63a14e9f0eaead", GitTreeState:"clean", BuildDate:"2018-03-26T16:55:54Z", GoVersion:"go1.9.3", Compiler:"gc", Platform:"linux/amd64"} Server Version: version.Info{Major:"1", Minor:"10", GitVersion:"v1.10.0", GitCommit:"fc32d2f3698e36b93322a3465f63a14e9f0eaead", GitTreeState:"clean", BuildDate:"2018-03-26T16:44:10Z", GoVersion:"go1.9.3", Compiler:"gc", Platform:"linux/amd64"}
nginx-ingress-controller:
cert-manager:
E0413 12:25:57.580259 1 controller.go:196] certificates controller: Re-queuing item "kube-system/therealhost.example.com" due to error processing: error waiting for key to be available for domain "therealhost.example.com": context deadline exceeded