Maybe the process is still responding on an active session? If you can't ping a host, that only means you cannot ping it. -----Original Message----- From: Steven Vacaroaia [mailto:stef97@xxxxxxxxx] Sent: woensdag 31 januari 2018 19:47 To: ceph-users Subject: Ceph - incorrect output of ceph osd tree Hi, Why is ceph osd tree reports that osd.4 is up when the server on which osd.4 is running is actually down ?? Any help will be appreciated [root@osd01 ~]# ping -c 2 osd02 PING osd02 (10.10.30.182) 56(84) bytes of data. >From osd01 (10.10.30.181) icmp_seq=1 Destination Host Unreachable From osd01 (10.10.30.181) icmp_seq=2 Destination Host Unreachable [root@osd01 ~]# ceph osd tree ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF -9 0 root ssds -10 0 host osd01-ssd -11 0 host osd02-ssd -12 0 host osd04-ssd -1 4.22031 root default -3 1.67967 host osd01 0 hdd 0.55989 osd.0 down 0 1.00000 3 hdd 0.55989 osd.3 down 0 1.00000 6 hdd 0.55989 osd.6 up 1.00000 1.00000 -5 1.67967 host osd02 1 hdd 0.55989 osd.1 down 1.00000 1.00000 4 hdd 0.55989 osd.4 up 1.00000 1.00000 7 hdd 0.55989 osd.7 down 1.00000 1.00000 -7 0.86096 host osd04 2 hdd 0.28699 osd.2 down 0 1.00000 5 hdd 0.28699 osd.5 down 1.00000 1.00000 8 hdd 0.28699 osd.8 down 1.00000 1.00000 [root@osd01 ~]# ceph tell osd.4 bench ^CError EINTR: problem getting command descriptions from osd.4 [root@osd01 ~]# ceph osd df ID CLASS WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS 0 hdd 0.55989 0 0 0 0 0 0 0 3 hdd 0.55989 0 0 0 0 0 0 0 6 hdd 0.55989 1.00000 573G 16474M 557G 2.81 0.84 0 1 hdd 0.55989 1.00000 573G 16516M 557G 2.81 0.84 0 4 hdd 0.55989 1.00000 573G 16465M 557G 2.80 0.84 0 7 hdd 0.55989 1.00000 573G 16473M 557G 2.81 0.84 0 2 hdd 0.28699 0 0 0 0 0 0 0 5 hdd 0.28699 1.00000 293G 16466M 277G 5.47 1.63 0 8 hdd 0.28699 1.00000 293G 16461M 277G 5.47 1.63 0 TOTAL 2881G 98857M 2784G 3.35 MIN/MAX VAR: 0.84/1.63 STDDEV: 1.30 [root@osd01 ~]# ceph osd df tree ID CLASS WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS TYPE NAME -9 0 - 0 0 0 0 0 - root ssds -10 0 - 0 0 0 0 0 - host osd01-ssd -11 0 - 0 0 0 0 0 - host osd02-ssd -12 0 - 0 0 0 0 0 - host osd04-ssd -1 4.22031 - 2881G 98857M 2784G 3.35 1.00 - root default -3 1.67967 - 573G 16474M 557G 2.81 0.84 - host osd01 0 hdd 0.55989 0 0 0 0 0 0 0 osd.0 3 hdd 0.55989 0 0 0 0 0 0 0 osd.3 6 hdd 0.55989 1.00000 573G 16474M 557G 2.81 0.84 0 osd.6 -5 1.67967 - 1720G 49454M 1671G 2.81 0.84 - host osd02 1 hdd 0.55989 1.00000 573G 16516M 557G 2.81 0.84 0 osd.1 4 hdd 0.55989 1.00000 573G 16465M 557G 2.80 0.84 0 osd.4 7 hdd 0.55989 1.00000 573G 16473M 557G 2.81 0.84 0 osd.7 -7 0.86096 - 587G 32928M 555G 5.47 1.63 - host osd04 2 hdd 0.28699 0 0 0 0 0 0 0 osd.2 5 hdd 0.28699 1.00000 293G 16466M 277G 5.47 1.63 0 osd.5 8 hdd 0.28699 1.00000 293G 16461M 277G 5.47 1.63 0 osd.8 TOTAL 2881G 98857M 2784G 3.35 MIN/MAX VAR: 0.84/1.63 STDDEV: 1.30 _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com