Closed pguerin3 closed 1 year ago
Correction - the last master node does install, so 3 masters in total are installed. But something is attempting to be installed after the 3rd master node.
Installing the Kubernetes module is failing. The likely cause is too little memory on your worker node.
If you’re using the vagrant-env
plug-in, copy your .env
file to .local.env
and increase the memory allocated for your WORKER_MEMORY
. If you don’t use the plug-in edit the Vagrantfile
directly (not recommended).
I suggest closing this issue. I can replicate by setting WORKER_MEMORY
in .env.local
to low amount.
What value did you use for the WORKER_MEMORY?
What value did you use for the WORKER_MEMORY?
To fail, usually less than 680
.
To work, at least 1024
. Though 2048
or higher would make more sense to run useful workload on your cluster and/or to test.
@pguerin3 - I just took a closer look at your screenshot. How exactly did you enable HA? In your screenshot, deploying the K8s modules is happening on master1
, which shouldn't be the case in HA setup.
To enable HA, simply set MULTI_MASTER=true
in .env.local
(assuming you have the vagrant-env
plugin). This will:
STANDALONE_OPERATOR=true
;operator
K8s modules will then be installed once the operator
VM is up and running.
I have now installed the plugin and created the following .env.local file:
MULTI_MASTER=true
WORKER_MEMORY=2048
SUBNET=192.168.56
Plus the new files mentioned in this issue: 468 I'll report back on the result soon.
Hi @pguerin3
Code is already merged. You can do a git clone
to get the latest.
I also sugguest you copy the full .env onto .env.local
so you can play around.
The error I get is:
Environment Host OS: Oracle Linux 9 with 24GB memory Kernel version (for Linux host): 5.14.0-162.6.1.el9_1.x86_64 Vagrant version: 2.3.4 Vagrant provider: Virtualbox 7.0.6r155176 Vagrant project: OLCNE
I would need more information to help. Would you: 1) Login to your operator node, and share the log file in question:
vagrant ssh operator
cat /var/tmp/cmd_OXQEP.log
2) Login to your master1
node and check with repo provides conmon
vagrant ssh master1
dnf list installed conmon
Also, I would recommend you set VERBOSE=true
in your .env.local
and try again:
vagrant destroy -f
vagrant up
(again, I'm assuming you have a current clone of this Git repo)
Finally I got the cluster up.
My issue is that a host with 24GB of memory is barely enough, so I had to find the right balance via trial and error for the VMs. My .env.local looks like this:
MULTI_MASTER=true
WORKER_MEMORY=1048
OPERATOR_MEMORY=3128
MASTER_MEMORY=5128
SUBNET=192.168.56
To do anything useful with the cluster, I would say that you need a host with at least 32GB.
I'm glad your issue is resolved. I run all my tests on my old Intel MacBook with 16GB of RAM, and you're right, more RAM would be nice.
Error installing with high availability enabled No problem, if I don't enable the high-availability feature:
But if I enable the feature (ie 3 master nodes), then the last master node will fail with:
Environment:
Additional information
Add any other information about the issue here (console log, ...).