concourse / hush-house

Concourse k8s-based environment
https://hush-house.pivotal.io
30 stars 23 forks source link

Road to a supported installation #27

Closed cirocosta closed 5 years ago

cirocosta commented 5 years ago

Hey,

Below is a list of items that we will need to complete/confirm before we can start moving/adding workloads onto hush house:

Thanks!


cc @scottietremendous

YoussB commented 5 years ago

after talking to @cirocosta this morning, the steps that we should take are as follows:

@scottietremendous wdyt?

YoussB commented 5 years ago

Extra things to think about: (not for hush-house GA but might be useful)

scottietremendous commented 5 years ago

@YoussB

I tend to think we should just go with the repo for incident reporting since that's what we basically do now with Wings.

On the second note, I think we should try out as many new features that helm can provide us as possible. I think it's important we still use this as a place to experiment.

aegershman commented 5 years ago

Decide between having the workloads on top of GKE or PKS at the moment, deploying hush-house in PKS wouldn't give the team much more data than we'd get from GKE, where we already have it running, allowing us to not have to learn any details of PKS and just move directly to what we already have.

Apologies for acting as some random dude interjecting my opinion here, but I'd gently suggest reconsidering.

Thoughts:

  1. There aren't too many details to learn that are unique to PKS's actual kubernetes clusters, but if you're going to be dog-fooding this with the intent for Pivotal customers to parrot (like me! :D) it wouldn't hurt to get first-hand experience on the toolset that they'll be using, what config options are available when setting up cluster plans, getting credentials to the cluster, etc.
  2. If nothing else, it would be beneficial to remove as much IaaS-specific implementation as possible. The current deployment appears to use IaaS/GKE-specific implementation details for the web/worker.nodeSelector and loadBalancerIP. It seems like there's not just an opportunity to experiment with Concourse's stability on kubernetes, but the operationalization of it's deployment
  3. Speaking of operationalization, you may also consider dogfooding running multiple Concourse environments like a "sandbox" to validate, then promoting those changes to production. It's beneficial because not only does it provide the practical value of multiple environments for testing changes, but it helps suss out optimizing workflows for handling "promoting" a series of changes in sandbox to prod.

3.1. Dogfooding the "environment promotion" workflow is beneficial because it helps drive out optimizations for declarative configuration that can be statically defined in a helm values.yml and not require an operator to go through "upgrade steps". The more things are operationalized such that configuration param keys: & values can be set and not require an operator to perform "in-between" steps, the better.

thanks for listening to my dissertation 👍

YoussB commented 5 years ago

Hey @aegershman,

thanks for the thoughtful comment. It makes a lot of sense 👍.

cirocosta commented 5 years ago

Hey @YoussB ,

In case we end up going with using namespaced secrets as a way of leveraging k8s cred mgmt, we'd need this one tackled first https://github.com/concourse/docs/issues/96 so that we can give a reference for the teams who end up consuming it.

Thanks!

YoussB commented 5 years ago

makes sense :+1: