crusmap show wrong osd for PGs (EC-Pool)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi all,
I had an issue on an hammer-cluster (0.94.9 - ugraded from 0.94.7 today).

There are three PGs incomplete:
root@ceph-06:~# ceph health detail
HEALTH_WARN 3 pgs incomplete; 3 pgs stuck inactive; 3 pgs stuck unclean
pg 24.cc is stuck inactive for 595902.285007, current state incomplete, last acting [8,111,12,88,128,44,56] pg 24.59 is stuck inactive for 596201.505599, current state remapped+incomplete, last acting [2147483647,77,55,2147483647,26,12,9] pg 16.142 is stuck inactive for 596297.260651, current state remapped+incomplete, last acting [74,12,2147483647,2147483647,5] pg 24.cc is stuck unclean for 599519.131339, current state incomplete, last acting [8,111,12,88,128,44,56] pg 24.59 is stuck unclean for 604346.612898, current state remapped+incomplete, last acting [2147483647,77,55,2147483647,26,12,9] pg 16.142 is stuck unclean for 662937.231308, current state remapped+incomplete, last acting [74,12,2147483647,2147483647,5] pg 16.142 is remapped+incomplete, acting [74,12,2147483647,2147483647,5] (reducing pool ecarchiv min_size from 3 may help; search ceph.com/docs for 'incomplete') pg 24.cc is incomplete, acting [8,111,12,88,128,44,56] (reducing pool ec7archiv min_size from 5 may help; search ceph.com/docs for 'incomplete') pg 24.59 is remapped+incomplete, acting [2147483647,77,55,2147483647,26,12,9] (reducing pool ec7archiv min_size from 5 may help; search ceph.com/docs for 'incomplete')

osd-12 was rebuildet some days ago - but the location of the other slices are not all right:

24.cc crushmap: [8,111,12,88,128,44,56]
real live: [8,121, X,88,130,44,56] - due the new osd-12 and the wrong searchlist (osd-121 + osd-130) the PG is undersized!

/var/lib/ceph/osd/ceph-8/current/24.ccs0_head
/var/lib/ceph/osd/ceph-44/current/24.ccs5_head
/var/lib/ceph/osd/ceph-56/current/24.ccs6_head
/var/lib/ceph/osd/ceph-88/current/24.ccs3_head
/var/lib/ceph/osd/ceph-121/current/24.ccs1_head
/var/lib/ceph/osd/ceph-130/current/24.ccs4_head

the same with the other PGs (two EC-5+2 + one EC-3+2)

/var/lib/ceph/osd/ceph-9/current/24.59s6_head
/var/lib/ceph/osd/ceph-26/current/24.59s4_head
/var/lib/ceph/osd/ceph-55/current/24.59s2_head
/var/lib/ceph/osd/ceph-77/current/24.59s1_head
/var/lib/ceph/osd/ceph-110/current/24.59s3_head
/var/lib/ceph/osd/ceph-122/current/24.59s0_head

/var/lib/ceph/osd/ceph-123/current/16.142s2_head
/var/lib/ceph/osd/ceph-5/current/16.142s4_head
/var/lib/ceph/osd/ceph-74/current/16.142s0_head
/var/lib/ceph/osd/ceph-115/current/16.142s3_head

set_choose_tries is set to 200.

ceph pg repair don't work.

Howto let calculate ceph the "right" osds?

Any hints?

Udo
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux