OSD wrongly marked up (well, half up)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi all,

On my test cluster I have 3 MON, 2 MDS and 2 OSD. I'm doing some
failover test on OSD and got a strange thing on the status.
The 2 nodes hosting the OSDs have been shutdown but the status continu
to 'see' one alive :

# ceph -v
ceph version 0.29 (commit:8e69c39f69936e2912a887247c6e268d1c9059ed)
# uname -a
Linux test2 2.6.38-8-server #42-Ubuntu SMP Mon Apr 11 03:49:04 UTC
2011 x86_64 x86_64 x86_64 GNU/Linux

root@test2:~# ceph health
2011-06-11 17:03:38.492734 mon <- [health]
2011-06-11 17:03:38.493913 mon1 -> 'HEALTH_WARN 594 pgs degraded,
551/1102 degraded (50.000%); 1/2 osds down, 1/2 osds out' (0)

root@test2:~# ceph osd stat
2011-06-11 17:03:48.071885 mon <- [osd,stat]
2011-06-11 17:03:48.073290 mon1 -> 'e31: 2 osds: 1 up, 1 in' (0)

root@test2:~# ceph mds stat
2011-06-11 17:03:54.868986 mon <- [mds,stat]
2011-06-11 17:03:54.870418 mon1 -> 'e48: 1/1/1 up {0=test4=up:active},
1 up:standby' (0)

root@test2:~# ceph mon stat
2011-06-11 17:04:09.638549 mon <- [mon,stat]
2011-06-11 17:04:09.639994 mon0 -> 'e1: 3 mons at
{0=10.1.56.231:6789/0,1=10.1.56.232:6789/0,2=10.1.56.233:6789/0},
election epoch 508, quorum 0,1,2' (0)

How could it be, is it a bug ?
(be sure I triple checked that my 2 osd nodes are really shutdown)

Thanks !
Wilfrid
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux