Re: Problems after increasing number of PGs in a pool

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I guess the pool is mapped to SSDs only from the name and you only got 20 SSDs.
So you should have about ~2000 effective PGs taking replication into account.

Your pool has ~10k effective PGs with k+m=5 and you seem to have 5
more pools....

Check "ceph osd df tree" to see how many PGs per OSD you got.

Try increasing these two options to "fix" it.

mon max pg per osd
osd max pg per osd hard ratio


Paul
Am Fr., 28. Sep. 2018 um 18:05 Uhr schrieb Vladimir Brik
<vladimir.brik@xxxxxxxxxxxxxxxx>:
>
> Hello
>
> I've attempted to increase the number of placement groups of the pools
> in our test cluster and now ceph status (below) is reporting problems. I
> am not sure what is going on or how to fix this. Troubleshooting
> scenarios in the docs don't seem to quite match what I am seeing.
>
> I have no idea how to begin to debug this. I see OSDs listed in
> "blocked_by" of pg dump, but don't know how to interpret that. Could
> somebody assist please?
>
> I attached output of "ceph pg dump_stuck -f json-pretty" just in case.
>
> The cluster consists of 5 hosts, each with 16 HDDs and 4 SSDs. I am
> running 13.2.2.
>
> This is the affected pool:
> pool 6 'fs-data-ec-ssd' erasure size 5 min_size 4 crush_rule 6
> object_hash rjenkins pg_num 2048 pgp_num 2048 last_change 2493 lfor
> 0/2491 flags hashpspool,ec_overwrites stripe_width 12288 application cephfs
>
>
> Thanks,
>
> Vlad
>
>
> ceph health
>
>   cluster:
>     id:     47caa1df-42be-444d-b603-02cad2a7fdd3
>     health: HEALTH_WARN
>             Reduced data availability: 155 pgs inactive, 47 pgs peering,
> 64 pgs stale
>             Degraded data redundancy: 321039/114913606 objects degraded
> (0.279%), 108 pgs degraded, 108 pgs undersized
>
>   services:
>     mon: 5 daemons, quorum ceph-1,ceph-2,ceph-3,ceph-4,ceph-5
>     mgr: ceph-3(active), standbys: ceph-2, ceph-5, ceph-1, ceph-4
>     mds: cephfs-1/1/1 up  {0=ceph-5=up:active}, 4 up:standby
>     osd: 100 osds: 100 up, 100 in; 165 remapped pgs
>
>   data:
>     pools:   6 pools, 5120 pgs
>     objects: 22.98 M objects, 88 TiB
>     usage:   154 TiB used, 574 TiB / 727 TiB avail
>     pgs:     3.027% pgs not active
>              321039/114913606 objects degraded (0.279%)
>              4903 active+clean
>              105  activating+undersized+degraded+remapped
>              61   stale+active+clean
>              47   remapped+peering
>              3    stale+activating+undersized+degraded+remapped
>              1    active+clean+scrubbing+deep
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



-- 
Paul Emmerich

Looking for help with your Ceph cluster? Contact us at https://croit.io

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux