Cephadm - Adding host to migrated cluster

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Greetings everyone,

 

We recently moved a ceph-ansible cluster running pacific on centos 8 to
centos 8 stream and then upgraded to quincy using cephadm after converting
to cephadm.  Everything with the transition worked but recently we decided
to add another node to the cluster with 10 more drives.  We were able to go
to the web interface and add the host ( with the IP and name ), which spun
up the basic management containers on the new node.  We then went to the OSD
section to add the drives which were showing as available.  They were all
recognized, so the drives were added via the web console.  Cephadm spun up
the OSDs and that's where things are stuck.  The OSDs show up in the cluster
but are out now.  They came up but were then marked down and later out.  We
purged them then zapped the drives and after about 10 minutes, cephadm had
added them back automatically.  It then did the same thing, showed them up,
then down and put them out.  When I look at "ceph osd tree", it shows the
drives but they don't show up under any host ( they are on host osdserver6
).  I am trying to figure out why they are not being put under a host since
the host server was added to cephadm and the server install checks with
cephadm were good.  The maintenance containers are running on the host, no
issues.  Any ideas would be greatly appreciated.

 

-16          36.38199      host osdserver5

20    ssd    3.63820          osd.20              up   1.00000  1.00000

22    ssd    3.63820          osd.22              up   1.00000  1.00000

23    ssd    3.63820          osd.23              up   1.00000  1.00000

24    ssd    3.63820          osd.24              up   1.00000  1.00000

44    ssd    3.63820          osd.44              up   1.00000  1.00000

45    ssd    3.63820          osd.45              up   1.00000  1.00000

46    ssd    3.63820          osd.46              up   1.00000  1.00000

47    ssd    3.63820          osd.47              up   1.00000  1.00000

48    ssd    3.63820          osd.48              up   1.00000  1.00000

49    ssd    3.63820          osd.49              up   1.00000  1.00000

37                 0  osd.37                    down   1.00000  1.00000

50                 0  osd.50                    down   1.00000  1.00000

51                 0  osd.51                    down   1.00000  1.00000

52                 0  osd.52                    down   1.00000  1.00000

53                 0  osd.53                    down   1.00000  1.00000

54                 0  osd.54                    down   1.00000  1.00000

55                 0  osd.55                    down   1.00000  1.00000

56                 0  osd.56                    down   1.00000  1.00000

57                 0  osd.57                    down   1.00000  1.00000

58                 0  osd.58                    down   1.00000  1.00000

 

 

Regards,

-Brent

 

Existing Clusters:

Test: Quincy 17.2.3 ( all virtual on nvme )

US Production(HDD): Octopus 15.2.16 with 11 osd servers, 3 mons, 4 gateways,
2 iscsi gateways

UK Production(HDD): Nautilus 14.2.22 with 18 osd servers, 3 mons, 4
gateways, 2 iscsi gateways

US Production(SSD): Quincy 17.2.3 Cephadm with 6 osd servers, 5 mons, 4
gateways, 2 iscsi gateways

UK Production(SSD): Quincy 17.2.3 with 6 osd servers, 5 mons, 4 gateways

 

 

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux