Hi all,
I had an issue on an hammer-cluster (0.94.9 - ugraded from 0.94.7
today).
There are three PGs incomplete:
root@ceph-06:~# ceph health detail
HEALTH_WARN 3 pgs incomplete; 3 pgs stuck inactive; 3 pgs stuck unclean
pg 24.cc is stuck inactive for 595902.285007, current state incomplete,
last acting [8,111,12,88,128,44,56]
pg 24.59 is stuck inactive for 596201.505599, current state
remapped+incomplete, last acting [2147483647,77,55,2147483647,26,12,9]
pg 16.142 is stuck inactive for 596297.260651, current state
remapped+incomplete, last acting [74,12,2147483647,2147483647,5]
pg 24.cc is stuck unclean for 599519.131339, current state incomplete,
last acting [8,111,12,88,128,44,56]
pg 24.59 is stuck unclean for 604346.612898, current state
remapped+incomplete, last acting [2147483647,77,55,2147483647,26,12,9]
pg 16.142 is stuck unclean for 662937.231308, current state
remapped+incomplete, last acting [74,12,2147483647,2147483647,5]
pg 16.142 is remapped+incomplete, acting [74,12,2147483647,2147483647,5]
(reducing pool ecarchiv min_size from 3 may help; search ceph.com/docs
for 'incomplete')
pg 24.cc is incomplete, acting [8,111,12,88,128,44,56] (reducing pool
ec7archiv min_size from 5 may help; search ceph.com/docs for
'incomplete')
pg 24.59 is remapped+incomplete, acting
[2147483647,77,55,2147483647,26,12,9] (reducing pool ec7archiv min_size
from 5 may help; search ceph.com/docs for 'incomplete')
osd-12 was rebuildet some days ago - but the location of the other
slices are not all right:
24.cc crushmap: [8,111,12,88,128,44,56]
real live: [8,121, X,88,130,44,56] - due the new osd-12 and the
wrong searchlist (osd-121 + osd-130) the PG is undersized!
/var/lib/ceph/osd/ceph-8/current/24.ccs0_head
/var/lib/ceph/osd/ceph-44/current/24.ccs5_head
/var/lib/ceph/osd/ceph-56/current/24.ccs6_head
/var/lib/ceph/osd/ceph-88/current/24.ccs3_head
/var/lib/ceph/osd/ceph-121/current/24.ccs1_head
/var/lib/ceph/osd/ceph-130/current/24.ccs4_head
the same with the other PGs (two EC-5+2 + one EC-3+2)
/var/lib/ceph/osd/ceph-9/current/24.59s6_head
/var/lib/ceph/osd/ceph-26/current/24.59s4_head
/var/lib/ceph/osd/ceph-55/current/24.59s2_head
/var/lib/ceph/osd/ceph-77/current/24.59s1_head
/var/lib/ceph/osd/ceph-110/current/24.59s3_head
/var/lib/ceph/osd/ceph-122/current/24.59s0_head
/var/lib/ceph/osd/ceph-123/current/16.142s2_head
/var/lib/ceph/osd/ceph-5/current/16.142s4_head
/var/lib/ceph/osd/ceph-74/current/16.142s0_head
/var/lib/ceph/osd/ceph-115/current/16.142s3_head
set_choose_tries is set to 200.
ceph pg repair don't work.
Howto let calculate ceph the "right" osds?
Any hints?
Udo
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com