batrick / ceph-linode

Launch Ceph using the Linode VPS provider
GNU General Public License v3.0
13 stars 10 forks source link

fatal: [mon-000]: FAILED! => #60

Closed selukov closed 3 years ago

selukov commented 3 years ago

I can't start ceph-linode, the installation breaks at shell: /root/cephadm bootstrap --allow-fqdn-hostname --mon-ip {{ monitor_address }}

TASK [bootstrap octopus] **********************************************************************************************************************************************************************************
fatal: [mon-000]: FAILED! => {
    "changed": true,
    "cmd": "/root/cephadm bootstrap --allow-fqdn-hostname --mon-ip 45.56.94.64",
    "delta": "0:12:02.194868",
    "end": "2020-09-23 13:40:01.039117",
    "rc": 1,
    "start": "2020-09-23 13:27:58.844249"
}

STDERR:

INFO:cephadm:Verifying podman|docker is present...
INFO:cephadm:Verifying lvm2 is present...
INFO:cephadm:Verifying time synchronization is in place...
INFO:cephadm:Unit chronyd.service is enabled and running
INFO:cephadm:Repeating the final host check...
INFO:cephadm:podman|docker (/usr/bin/podman) is present
INFO:cephadm:systemctl is present
INFO:cephadm:lvcreate is present
INFO:cephadm:Unit chronyd.service is enabled and running
INFO:cephadm:Host looks OK
INFO:root:Cluster fsid: 98f5f4ec-fda0-11ea-8836-f23c922d25b7
INFO:cephadm:Verifying IP 45.56.94.64 port 3300 ...
INFO:cephadm:Verifying IP 45.56.94.64 port 6789 ...
INFO:cephadm:Mon IP 45.56.94.64 is in CIDR network 45.56.94.0/24
INFO:cephadm:Pulling container image ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64...
INFO:cephadm:Extracting ceph user uid/gid from container image...
INFO:cephadm:Creating initial keys...
INFO:cephadm:Creating initial monmap...
INFO:cephadm:Creating mon...
INFO:cephadm:Waiting for mon to start...
INFO:cephadm:Waiting for mon...
INFO:cephadm:/usr/bin/ceph:timeout after 60 seconds
INFO:cephadm:Non-zero exit code -9 from /usr/bin/podman run --rm --net=host --ipc=host -e CONTAINER_IMAGE=ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64 -e NODE_NAME=li896-64 -v /var/lib/ceph/98f5f4ec-fda0-11ea-8836-f23c922d25b7/mon.li896-64:/var/lib/ceph/mon/ceph-li896-64:z -v /tmp/ceph-tmpt1gd66zz:/etc/ceph/ceph.client.admin.keyring:z -v /tmp/ceph-tmp3w_b573j:/etc/ceph/ceph.conf:z --entrypoint /usr/bin/ceph ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64 status
INFO:cephadm:mon not available, waiting (1/10)...
INFO:cephadm:/usr/bin/ceph:timeout after 60 seconds
INFO:cephadm:Non-zero exit code -9 from /usr/bin/podman run --rm --net=host --ipc=host -e CONTAINER_IMAGE=ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64 -e NODE_NAME=li896-64 -v /var/lib/ceph/98f5f4ec-fda0-11ea-8836-f23c922d25b7/mon.li896-64:/var/lib/ceph/mon/ceph-li896-64:z -v /tmp/ceph-tmpt1gd66zz:/etc/ceph/ceph.client.admin.keyring:z -v /tmp/ceph-tmp3w_b573j:/etc/ceph/ceph.conf:z --entrypoint /usr/bin/ceph ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64 status
INFO:cephadm:mon not available, waiting (2/10)...
INFO:cephadm:/usr/bin/ceph:timeout after 60 seconds
INFO:cephadm:Non-zero exit code -9 from /usr/bin/podman run --rm --net=host --ipc=host -e CONTAINER_IMAGE=ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64 -e NODE_NAME=li896-64 -v /var/lib/ceph/98f5f4ec-fda0-11ea-8836-f23c922d25b7/mon.li896-64:/var/lib/ceph/mon/ceph-li896-64:z -v /tmp/ceph-tmpt1gd66zz:/etc/ceph/ceph.client.admin.keyring:z -v /tmp/ceph-tmp3w_b573j:/etc/ceph/ceph.conf:z --entrypoint /usr/bin/ceph ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64 status
INFO:cephadm:mon not available, waiting (3/10)...
INFO:cephadm:/usr/bin/ceph:timeout after 60 seconds
INFO:cephadm:Non-zero exit code -9 from /usr/bin/podman run --rm --net=host --ipc=host -e CONTAINER_IMAGE=ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64 -e NODE_NAME=li896-64 -v /var/lib/ceph/98f5f4ec-fda0-11ea-8836-f23c922d25b7/mon.li896-64:/var/lib/ceph/mon/ceph-li896-64:z -v /tmp/ceph-tmpt1gd66zz:/etc/ceph/ceph.client.admin.keyring:z -v /tmp/ceph-tmp3w_b573j:/etc/ceph/ceph.conf:z --entrypoint /usr/bin/ceph ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64 status
INFO:cephadm:mon not available, waiting (4/10)...
INFO:cephadm:/usr/bin/ceph:timeout after 60 seconds
INFO:cephadm:Non-zero exit code -9 from /usr/bin/podman run --rm --net=host --ipc=host -e CONTAINER_IMAGE=ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64 -e NODE_NAME=li896-64 -v /var/lib/ceph/98f5f4ec-fda0-11ea-8836-f23c922d25b7/mon.li896-64:/var/lib/ceph/mon/ceph-li896-64:z -v /tmp/ceph-tmpt1gd66zz:/etc/ceph/ceph.client.admin.keyring:z -v /tmp/ceph-tmp3w_b573j:/etc/ceph/ceph.conf:z --entrypoint /usr/bin/ceph ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64 status
INFO:cephadm:mon not available, waiting (5/10)...
INFO:cephadm:/usr/bin/ceph:timeout after 60 seconds
INFO:cephadm:Non-zero exit code -9 from /usr/bin/podman run --rm --net=host --ipc=host -e CONTAINER_IMAGE=ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64 -e NODE_NAME=li896-64 -v /var/lib/ceph/98f5f4ec-fda0-11ea-8836-f23c922d25b7/mon.li896-64:/var/lib/ceph/mon/ceph-li896-64:z -v /tmp/ceph-tmpt1gd66zz:/etc/ceph/ceph.client.admin.keyring:z -v /tmp/ceph-tmp3w_b573j:/etc/ceph/ceph.conf:z --entrypoint /usr/bin/ceph ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64 status
INFO:cephadm:mon not available, waiting (6/10)...
INFO:cephadm:/usr/bin/ceph:timeout after 60 seconds
INFO:cephadm:Non-zero exit code -9 from /usr/bin/podman run --rm --net=host --ipc=host -e CONTAINER_IMAGE=ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64 -e NODE_NAME=li896-64 -v /var/lib/ceph/98f5f4ec-fda0-11ea-8836-f23c922d25b7/mon.li896-64:/var/lib/ceph/mon/ceph-li896-64:z -v /tmp/ceph-tmpt1gd66zz:/etc/ceph/ceph.client.admin.keyring:z -v /tmp/ceph-tmp3w_b573j:/etc/ceph/ceph.conf:z --entrypoint /usr/bin/ceph ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64 status
INFO:cephadm:mon not available, waiting (7/10)...
INFO:cephadm:/usr/bin/ceph:timeout after 60 seconds
INFO:cephadm:Non-zero exit code -9 from /usr/bin/podman run --rm --net=host --ipc=host -e CONTAINER_IMAGE=ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64 -e NODE_NAME=li896-64 -v /var/lib/ceph/98f5f4ec-fda0-11ea-8836-f23c922d25b7/mon.li896-64:/var/lib/ceph/mon/ceph-li896-64:z -v /tmp/ceph-tmpt1gd66zz:/etc/ceph/ceph.client.admin.keyring:z -v /tmp/ceph-tmp3w_b573j:/etc/ceph/ceph.conf:z --entrypoint /usr/bin/ceph ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64 status
INFO:cephadm:mon not available, waiting (8/10)...
INFO:cephadm:/usr/bin/ceph:timeout after 60 seconds
INFO:cephadm:Non-zero exit code -9 from /usr/bin/podman run --rm --net=host --ipc=host -e CONTAINER_IMAGE=ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64 -e NODE_NAME=li896-64 -v /var/lib/ceph/98f5f4ec-fda0-11ea-8836-f23c922d25b7/mon.li896-64:/var/lib/ceph/mon/ceph-li896-64:z -v /tmp/ceph-tmpt1gd66zz:/etc/ceph/ceph.client.admin.keyring:z -v /tmp/ceph-tmp3w_b573j:/etc/ceph/ceph.conf:z --entrypoint /usr/bin/ceph ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64 status
INFO:cephadm:mon not available, waiting (9/10)...
INFO:cephadm:/usr/bin/ceph:timeout after 60 seconds
INFO:cephadm:Non-zero exit code -9 from /usr/bin/podman run --rm --net=host --ipc=host -e CONTAINER_IMAGE=ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64 -e NODE_NAME=li896-64 -v /var/lib/ceph/98f5f4ec-fda0-11ea-8836-f23c922d25b7/mon.li896-64:/var/lib/ceph/mon/ceph-li896-64:z -v /tmp/ceph-tmpt1gd66zz:/etc/ceph/ceph.client.admin.keyring:z -v /tmp/ceph-tmp3w_b573j:/etc/ceph/ceph.conf:z --entrypoint /usr/bin/ceph ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64 status
INFO:cephadm:mon not available, waiting (10/10)...
INFO:cephadm:/usr/bin/ceph:timeout after 60 seconds
INFO:cephadm:Non-zero exit code -9 from /usr/bin/podman run --rm --net=host --ipc=host -e CONTAINER_IMAGE=ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64 -e NODE_NAME=li896-64 -v /var/lib/ceph/98f5f4ec-fda0-11ea-8836-f23c922d25b7/mon.li896-64:/var/lib/ceph/mon/ceph-li896-64:z -v /tmp/ceph-tmpt1gd66zz:/etc/ceph/ceph.client.admin.keyring:z -v /tmp/ceph-tmp3w_b573j:/etc/ceph/ceph.conf:z --entrypoint /usr/bin/ceph ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64 status
ERROR: mon not available after 10 tries

MSG:

non-zero return code

PLAY RECAP ************************************************************************************************************************************************************************************************
client-000                 : ok=33   changed=27   unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   
mds-000                    : ok=34   changed=27   unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   
mgr-000                    : ok=34   changed=28   unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   
mon-000                    : ok=37   changed=28   unreachable=0    failed=1    skipped=1    rescued=0    ignored=0   
mon-001                    : ok=33   changed=26   unreachable=0    failed=0    skipped=1    rescued=0    ignored=0   
mon-002                    : ok=33   changed=26   unreachable=0    failed=0    skipped=1    rescued=0    ignored=0   
osd-000                    : ok=36   changed=29   unreachable=0    failed=0    skipped=0    rescued=0    ignored=1   
osd-001                    : ok=36   changed=29   unreachable=0    failed=0    skipped=0    rescued=0    ignored=1   
osd-002                    : ok=36   changed=29   unreachable=0    failed=0    skipped=0    rescued=0    ignored=1   

real    21m29,087s
user    1m46,031s
sys 0m39,729s

I replaced it to the stable version, for the developer version exactly the same error

  - set_fact:
      CEPHADM_IMAGE: ceph/daemon-base:v5.0.3-stable-5.0-octopus-centos-8-x86_64
      CEPHADM_REPO: --release octopus

Ceph tries to start on mon-000 but I don’t understand why it doesn’t work

./ansible-ssh mon-000
ps -ax
   4619 ?        S      0:00 (sd-pam)
  33672 ?        I      0:00 [kworker/u2:0-events_unbound]
  36859 ?        I      0:00 [kworker/0:2-events_power_efficient]
  36935 ?        S      0:00 /usr/sbin/chronyd
  38931 ?        R      0:00 [kworker/u2:1-flush-8:0]
  41421 ?        Ss     0:00 /bin/bash /var/lib/ceph/98f5f4ec-fda0-11ea-8836-f23c922d25b7/mon.li896-64/unit.run
  41450 ?        Sl     0:00 /usr/bin/podman run --rm --net=host --ipc=host --privileged --group-add=disk --name ceph-98f5f4ec-fda0-11ea-8836-f23c922d25b7-mon.li896-64 -e CONTAINER_IMAGE=ceph/daemon-base
  41458 ?        Ssl    0:00 /usr/bin/conmon --api-version 1 -s -c e3c4e602d08452830cfb768ecb187da134f160e73def334b1481e926ae0c884e -u e3c4e602d08452830cfb768ecb187da134f160e73def334b1481e926ae0c884e -r 
  41469 ?        Ssl    0:00 /usr/libexec/platform-python -s /usr/bin/ceph status
  41479 ?        Ssl    0:00 /usr/bin/conmon --api-version 1 -s -c 8438619ccb7b3b44503d01367da715e67e0bf07c422b642e139893d25fba4994 -u 8438619ccb7b3b44503d01367da715e67e0bf07c422b642e139893d25fba4994 -r 
  41491 ?        Ssl    0:00 /usr/bin/ceph-mon -n mon.li896-64 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug  --default-mon-
  41576 ?        Ssl    0:00 /usr/bin/conmon --api-version 1 -s -c 6a6d696288002676da7b66ea2acf9edb2a065f7075c506bc18a7bf0c483f66d6 -u 6a6d696288002676da7b66ea2acf9edb2a065f7075c506bc18a7bf0c483f66d6 -r 
  41587 ?        Ssl    0:00 /usr/libexec/platform-python -s /usr/bin/ceph status
  41629 ?        Ssl    0:00 /usr/bin/conmon --api-version 1 -s -c 6ae2ec2c90d62c2e0187fdf5251f14a1bc639189e9cf913e03ee60ad91a1a85c -u 6ae2ec2c90d62c2e0187fdf5251f14a1bc639189e9cf913e03ee60ad91a1a85c -r 
  41640 ?        Ssl    0:00 /usr/libexec/platform-python -s /usr/bin/ceph status
  41674 ?        Ss     0:00 sshd: root [priv]
  41677 ?        S      0:00 sshd: root@pts/1
  41678 pts/1    Ss     0:00 -bash
  41708 ?        Ssl    0:00 /usr/bin/conmon --api-version 1 -s -c 5581995d13c14704d35b7facf0212bad115dcc965928d448f7cc3b7ca9d84139 -u 5581995d13c14704d35b7facf0212bad115dcc965928d448f7cc3b7ca9d84139 -r 
  41719 ?        Ssl    0:00 /usr/libexec/platform-python -s /usr/bin/ceph status
  41761 ?        Ssl    0:00 /usr/bin/conmon --api-version 1 -s -c 02dba76f865472736a1631b5456f0dcc71cf0222bf58f37e1941f5b3545955a0 -u 02dba76f865472736a1631b5456f0dcc71cf0222bf58f37e1941f5b3545955a0 -r 
  41772 ?        Ssl    0:00 /usr/libexec/platform-python -s /usr/bin/ceph status
  41813 ?        Ssl    0:00 /usr/bin/conmon --api-version 1 -s -c 0cb2a8f64f2b694a0a97bddd3c710c12fbf2fcad1f7a4c9f8432cd7d2fc44e3c -u 0cb2a8f64f2b694a0a97bddd3c710c12fbf2fcad1f7a4c9f8432cd7d2fc44e3c -r 
  41824 ?        Ssl    0:00 /usr/libexec/platform-python -s /usr/bin/ceph status
  41867 ?        Ssl    0:00 /usr/bin/conmon --api-version 1 -s -c ba18819bb3b978210d85015f9cf9d6bf9ab1fc949c09abf77d79c356ce443ac1 -u ba18819bb3b978210d85015f9cf9d6bf9ab1fc949c09abf77d79c356ce443ac1 -r 
  41878 ?        Ssl    0:00 /usr/libexec/platform-python -s /usr/bin/ceph status
  41924 ?        Ssl    0:00 /usr/bin/conmon --api-version 1 -s -c 40a2e69fed8781074f836f56806bee90a0036da0f9b8a5a268c6e4c54815a8f4 -u 40a2e69fed8781074f836f56806bee90a0036da0f9b8a5a268c6e4c54815a8f4 -r 
  41935 ?        Ssl    0:00 /usr/libexec/platform-python -s /usr/bin/ceph status
  41965 ?        I      0:00 [kworker/0:0-events]
  41979 ?        Ssl    0:00 /usr/bin/conmon --api-version 1 -s -c c3b0ed74fc6f64a1c525418efd0a7ad60b48d4283c36f3be96323e00bc69cc63 -u c3b0ed74fc6f64a1c525418efd0a7ad60b48d4283c36f3be96323e00bc69cc63 -r 
  41990 ?        Ssl    0:00 /usr/libexec/platform-python -s /usr/bin/ceph status
  42033 ?        Ssl    0:00 /usr/bin/conmon --api-version 1 -s -c 4a58314c03aa1793868dd4b68153dbc1fbeb75654c298dd6cdc8aa169356a77c -u 4a58314c03aa1793868dd4b68153dbc1fbeb75654c298dd6cdc8aa169356a77c -r 
  42044 ?        Ssl    0:00 /usr/libexec/platform-python -s /usr/bin/ceph status
  42087 ?        Ssl    0:00 /usr/bin/conmon --api-version 1 -s -c 8e202457a962fbd2a399b60abd98b6007c311a59b88605642860567513658d98 -u 8e202457a962fbd2a399b60abd98b6007c311a59b88605642860567513658d98 -r 
  42098 ?        Ssl    0:00 /usr/libexec/platform-python -s /usr/bin/ceph status
  42252 ?        I      0:00 [kworker/0:5-cgroup_destroy]
  42347 ?        I      0:00 [kworker/0:1-events]
  42423 ?        I      0:00 [kworker/0:4-events_power_efficient]
  42431 ?        Ss     0:00 /usr/sbin/anacron -s
  42434 ?        Ss     0:00 sshd: unknown [priv]
  42435 ?        S      0:00 sshd: unknown [net]
  42436 pts/1    R+     0:00 ps -ax
batrick commented 3 years ago
"cmd": "/root/cephadm bootstrap --allow-fqdn-hostname --mon-ip 45.56.94.64",

Normally this is a private IP. Did you change something? I just tried a fresh setup with the defaults and the cluster.json.sample and got:

TASK [bootstrap octopus] *****************************************************************************************************************************************************
changed: [mon-000] => {
    "changed": true,
    "cmd": "/root/cephadm bootstrap --allow-fqdn-hostname --mon-ip 192.168.215.25",
    "delta": "0:01:05.271916",
    "end": "2020-09-23 15:57:32.542617",
    "rc": 0,
    "start": "2020-09-23 15:56:27.270701"
}

STDERR:

INFO:cephadm:Verifying podman|docker is present...
INFO:cephadm:Verifying lvm2 is present...
INFO:cephadm:Verifying time synchronization is in place...
INFO:cephadm:Unit chronyd.service is enabled and running
INFO:cephadm:Repeating the final host check...
INFO:cephadm:podman|docker (/usr/bin/podman) is present
INFO:cephadm:systemctl is present
INFO:cephadm:lvcreate is present
INFO:cephadm:Unit chronyd.service is enabled and running
INFO:cephadm:Host looks OK
INFO:root:Cluster fsid: 56bdb28a-fdb5-11ea-b5f3-f23c922dbbf0
INFO:cephadm:Verifying IP 192.168.215.25 port 3300 ...
INFO:cephadm:Verifying IP 192.168.215.25 port 6789 ...
INFO:cephadm:Mon IP 192.168.215.25 is in CIDR network 192.168.128.0/17
INFO:cephadm:Pulling container image ceph/daemon-base:latest-octopus-devel...
INFO:cephadm:Extracting ceph user uid/gid from container image...
INFO:cephadm:Creating initial keys...
INFO:cephadm:Creating initial monmap...
INFO:cephadm:Creating mon...
INFO:cephadm:Waiting for mon to start...
INFO:cephadm:Waiting for mon...
INFO:cephadm:mon is available
INFO:cephadm:Assimilating anything we can from ceph.conf...
INFO:cephadm:Generating new minimal ceph.conf...
INFO:cephadm:Restarting the monitor...
INFO:cephadm:Setting mon public_network...
INFO:cephadm:Creating mgr...
INFO:cephadm:Verifying port 9283 ...
INFO:cephadm:Wrote keyring to /etc/ceph/ceph.client.admin.keyring
INFO:cephadm:Wrote config to /etc/ceph/ceph.conf
INFO:cephadm:Waiting for mgr to start...
INFO:cephadm:Waiting for mgr...
INFO:cephadm:mgr not available, waiting (1/10)...
INFO:cephadm:mgr not available, waiting (2/10)...
INFO:cephadm:mgr not available, waiting (3/10)...
INFO:cephadm:mgr not available, waiting (4/10)...
INFO:cephadm:mgr is available
INFO:cephadm:Enabling cephadm module...
INFO:cephadm:Waiting for the mgr to restart...
INFO:cephadm:Waiting for Mgr epoch 5...
INFO:cephadm:Mgr epoch 5 is available
INFO:cephadm:Setting orchestrator backend to cephadm...
INFO:cephadm:Generating ssh key...
INFO:cephadm:Wrote public SSH key to to /etc/ceph/ceph.pub
INFO:cephadm:Adding key to root@localhost's authorized_keys...
INFO:cephadm:Adding host li156-119.members.linode.com...
INFO:cephadm:Deploying mon service with default placement...
INFO:cephadm:Deploying mgr service with default placement...
INFO:cephadm:Deploying crash service with default placement...
INFO:cephadm:Enabling mgr prometheus module...
INFO:cephadm:Deploying prometheus service with default placement...
INFO:cephadm:Deploying grafana service with default placement...
INFO:cephadm:Deploying node-exporter service with default placement...
INFO:cephadm:Deploying alertmanager service with default placement...
INFO:cephadm:Enabling the dashboard module...
INFO:cephadm:Waiting for the mgr to restart...
INFO:cephadm:Waiting for Mgr epoch 13...
INFO:cephadm:Mgr epoch 13 is available
INFO:cephadm:Generating a dashboard self-signed certificate...
INFO:cephadm:Creating initial admin user...
INFO:cephadm:Fetching dashboard port number...
INFO:cephadm:Ceph Dashboard is now available at:

             URL: https://li156-119.members.linode.com:8443/
            User: admin
        Password: ttwmylt2m8

INFO:cephadm:You can access the Ceph CLI with:

        sudo /root/cephadm shell --fsid 56bdb28a-fdb5-11ea-b5f3-f23c922dbbf0 -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring

INFO:cephadm:Please consider enabling telemetry to help improve Ceph:

        ceph telemetry on

For more information see:

        https://docs.ceph.com/docs/master/mgr/telemetry/

INFO:cephadm:Bootstrap complete.
selukov commented 3 years ago

Thanks for meeting me on such short notice. I used cluster.json.sample, but it looks like there is no private IP in the systems

[root@li1006-182 ~]# ifconfig 
eth0: flags=4163<UP,BROADCAST,RUNNING,MULTICAST>  mtu 1500
        inet 45.33.56.182  netmask 255.255.255.0  broadcast 45.33.56.255
        inet6 fe80::f03c:92ff:fe2d:b773  prefixlen 64  scopeid 0x20<link>
        inet6 2600:3c01::f03c:92ff:fe2d:b773  prefixlen 64  scopeid 0x0<global>
        ether f2:3c:92:2d:b7:73  txqueuelen 1000  (Ethernet)
        RX packets 58399  bytes 531781966 (507.1 MiB)
        RX errors 0  dropped 66  overruns 0  frame 0
        TX packets 34765  bytes 2870409 (2.7 MiB)
        TX errors 0  dropped 0 overruns 0  carrier 0  collisions 0

lo: flags=73<UP,LOOPBACK,RUNNING>  mtu 65536
        inet 127.0.0.1  netmask 255.0.0.0
        inet6 ::1  prefixlen 128  scopeid 0x10<host>
        loop  txqueuelen 1000  (Local Loopback)
        RX packets 2651  bytes 159060 (155.3 KiB)
        RX errors 0  dropped 0  overruns 0  frame 0
        TX packets 2651  bytes 159060 (155.3 KiB)
        TX errors 0  dropped 0 overruns 0  carrier 0  collisions 0
batrick commented 3 years ago

Ah, I think the problem is that you're running ceph-linode locally and not in a linode VM (a bounce box). I have not recently tested that but it would explain why the IP address is wrong.

Can you try again with a nano-linode (the 1GB one) running ceph-linode?

selukov commented 3 years ago

I made two mistakes:

Ceph-linode installation was successful, Thank you

TASK [install octopus] **********************************************************************************************************************************************************************************************
changed: [client-000] => {
    "changed": true,
    "cmd": "/root/cephadm install ceph-common ceph-fuse",
    "delta": "0:00:17.679857",
    "end": "2020-09-24 10:44:39.809354",
    "rc": 0,
    "start": "2020-09-24 10:44:22.129497"
}

STDERR:

INFO:cephadm:Installing packages ['ceph-common', 'ceph-fuse']...

PLAY RECAP **********************************************************************************************************************************************************************************************************
client-000                 : ok=38   changed=30   unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   
mds-000                    : ok=41   changed=32   unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   
mgr-000                    : ok=38   changed=31   unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   
mon-000                    : ok=61   changed=46   unreachable=0    failed=0    skipped=1    rescued=0    ignored=0   
mon-001                    : ok=40   changed=31   unreachable=0    failed=0    skipped=1    rescued=0    ignored=0   
mon-002                    : ok=40   changed=31   unreachable=0    failed=0    skipped=1    rescued=0    ignored=0   
osd-000                    : ok=43   changed=34   unreachable=0    failed=0    skipped=0    rescued=0    ignored=1   
osd-001                    : ok=43   changed=34   unreachable=0    failed=0    skipped=0    rescued=0    ignored=1   
osd-002                    : ok=43   changed=34   unreachable=0    failed=0    skipped=0    rescued=0    ignored=1   

real    11m33.490s
user    1m32.022s
sys 0m35.319s
[root@li2158-77 ceph-linode]# python3 linode.py destroy
batrick commented 3 years ago

Good to hear you got it work. I will leave this open as I do need to fix ceph-linode to work when running locally.