Open colstrom opened 7 years ago
https://www.mail-archive.com/smartos-discuss@lists.smartos.org/msg02833.html references what appears to be the same issue. I will be increasing the available swap (currently 128GB per node, equal to RAM) to see if that sidesteps the issue.
When provisioning a KVM VM on my Triton deployment, Compute Nodes can reach a state where sdc-designation incorrectly considers it eligible for a VM that requires more RAM than the node has available.
Prior to provisioning, the available memory on this node is less than what is reported in the AdminUI (which reports 50.9GB available for this node).
Creating a new VM with
triton instance create
(or the AdminUI), specifying a package that requires MORE than what the AdminUI reports will ignore this host, and provision elsewhere. Specifying a package that requires less than the perceived capacity, but more than the actual capacity as reported bymdb
, leads to this failure.Workflow reports the job as
Running
, though the failures occur almost immediately. After approximately 20 minutes, the job is markedFailed
, and the VM is not destroyed (leaving resources allocated).Relevant log snippets are below. I can reproduce this easily, so if additional details would be helpful, please let me know.
/zones/4301b6cb-a382-6e6a-aa8c-e7fe04d88796/root/tmp/vm.log*
/var/svc/log/system-smartdc-vmadmd:default.log
The logs repeat like that for some time. Checking memory on the host, while that is happening yields the following:
for reference, the logs are similar when deliberately provisioning a VM that exceeds a node's available memory. In this case, 64GB requested vs 50.9 reported available, manually allocating the VM to a node that obviously cannot fit it).
/zones/d7bf4c1b-ac00-cee8-ff82-886a64c34f6a/root/tmp/vm.log*
/var/svc/log/system-smartdc-vmadmd:default.log