oracle / vagrant-projects

Vagrant projects for Oracle products and other examples
Universal Permissive License v1.0
946 stars 477 forks source link

Could not install module error #454

Closed pguerin3 closed 1 year ago

pguerin3 commented 2 years ago

Error installing with high availability enabled No problem, if I don't enable the high-availability feature: image

But if I enable the feature (ie 3 master nodes), then the last master node will fail with: image

Environment:

Additional information

Add any other information about the issue here (console log, ...).

pguerin3 commented 2 years ago

Correction - the last master node does install, so 3 masters in total are installed. But something is attempting to be installed after the 3rd master node.

hussam-qasem commented 2 years ago

Installing the Kubernetes module is failing. The likely cause is too little memory on your worker node.

If you’re using the vagrant-env plug-in, copy your .env file to .local.env and increase the memory allocated for your WORKER_MEMORY. If you don’t use the plug-in edit the Vagrantfile directly (not recommended).

hussam-qasem commented 1 year ago

I suggest closing this issue. I can replicate by setting WORKER_MEMORY in .env.local to low amount.

pguerin3 commented 1 year ago

What value did you use for the WORKER_MEMORY?

hussam-qasem commented 1 year ago

What value did you use for the WORKER_MEMORY?

To fail, usually less than 680.

To work, at least 1024. Though 2048 or higher would make more sense to run useful workload on your cluster and/or to test.

hussam-qasem commented 1 year ago

@pguerin3 - I just took a closer look at your screenshot. How exactly did you enable HA? In your screenshot, deploying the K8s modules is happening on master1, which shouldn't be the case in HA setup.

To enable HA, simply set MULTI_MASTER=true in .env.local (assuming you have the vagrant-env plugin). This will:

K8s modules will then be installed once the operator VM is up and running.

pguerin3 commented 1 year ago

I have now installed the plugin and created the following .env.local file:

MULTI_MASTER=true
WORKER_MEMORY=2048
SUBNET=192.168.56

Plus the new files mentioned in this issue: 468 I'll report back on the result soon.

hussam-qasem commented 1 year ago

Hi @pguerin3 Code is already merged. You can do a git clone to get the latest. I also sugguest you copy the full .env onto .env.local so you can play around.

pguerin3 commented 1 year ago

The error I get is: image

Environment Host OS: Oracle Linux 9 with 24GB memory Kernel version (for Linux host): 5.14.0-162.6.1.el9_1.x86_64 Vagrant version: 2.3.4 Vagrant provider: Virtualbox 7.0.6r155176 Vagrant project: OLCNE

hussam-qasem commented 1 year ago

I would need more information to help. Would you: 1) Login to your operator node, and share the log file in question:

vagrant ssh operator
cat /var/tmp/cmd_OXQEP.log

2) Login to your master1 node and check with repo provides conmon

vagrant ssh master1
dnf list installed conmon

Also, I would recommend you set VERBOSE=true in your .env.local and try again:

vagrant destroy -f
vagrant up

(again, I'm assuming you have a current clone of this Git repo)

pguerin3 commented 1 year ago

Finally I got the cluster up. image

My issue is that a host with 24GB of memory is barely enough, so I had to find the right balance via trial and error for the VMs. My .env.local looks like this:

MULTI_MASTER=true
WORKER_MEMORY=1048
OPERATOR_MEMORY=3128
MASTER_MEMORY=5128
SUBNET=192.168.56

To do anything useful with the cluster, I would say that you need a host with at least 32GB.

hussam-qasem commented 1 year ago

I'm glad your issue is resolved. I run all my tests on my old Intel MacBook with 16GB of RAM, and you're right, more RAM would be nice.