Too many PGs during filestore=>bluestore migration

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

I'm migrating my OSDs to bluestore (Luminous 12.2.10) recreating the
OSDs, everything looks good (just a few OSDs left), but yesterday a
weird thing happened after I set a wrong weigth to a newly migrate
OSD: should be 2 but I put 6 (hardcoded in my salt state, oops). I got
slow requests, when I realized that I stop this OSD.

I found things like this on my logs:

osd.125 492578 maybe_wait_for_max_pg withhold creation of pg 5.95e: 750 >= 750

So I destroyed this OSD again (when I realized that weight was 6) then
recreated it with 2. Things OK again...

Then I continued to migrate the OSDs, but after a few ones on this
same node I got slow requests again (all weights is 2 now). This time
slow req stopped only when I out all recent migrated OSDs on this
node.

Looking why I found the same kind of log, then I found what I don't
understand:

# ceph osd pool get <pool> pg_num
pg_num: 4096

*but* counting the PGs (using ceph osd df or mgr dash) I have 8964!

Why? Obviously this should not happened.

This OSD is part of a group of 3 nodes that serve only a small
specialized pool (not part of my main/large cluster). As I'm using
only 4TB I can easily migrate data to another pool then completely
destroy (and recreate) this buggy pool.

Anyway I just want to understand how/why that happened. Any clues?

--
Herbert
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux