Re: Fwd: HEALTH_WARN pool vol has too few pgs

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Yes, if I change the pg_num on current pool, cluster rebalance start..
Alternatively - I plan to do as below:
1. Createa a new pool with max possible pg_num (as per the pg calc).
2. Copy the current pool to new pool  (during this step - IO should be stopped)
3. Rename the curent pool current.old and rename the new pool to current pool.

After 3rd step - I guess, cluster should be fine without rebalance.

Thanks
Swami


On Thu, Feb 4, 2016 at 11:38 AM, Somnath Roy <Somnath.Roy@xxxxxxxxxxx> wrote:
> You can increase it, but, that will trigger rebalancing and based on the amount of data it will take some time before cluster is coming into clean state.
> Client IO performance will be affected during this.
> BTW this is not really an error , it is a warning because performance on that pool will be affected because of low pg count.
>
> Thanks & Regards
> Somnath
> -----Original Message-----
> From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf Of M Ranga Swami Reddy
> Sent: Wednesday, February 03, 2016 9:48 PM
> To: Ferhat Ozkasgarli
> Cc: ceph-users
> Subject: Re:  Fwd: HEALTH_WARN pool vol has too few pgs
>
> Current pg_num: 4096.  As per the PG num formula, no OSD * 100/pool size ->
> 184 * 100/3 = 6133, so I can increase to 8192. Is this solves the problem?
>
> Thanks
> Swami
>
> On Thu, Feb 4, 2016 at 2:14 AM, Ferhat Ozkasgarli <ozkasgarli@xxxxxxxxx> wrote:
>> As message satates, you must increase placement group number for the pool.
>> Because 108T data require larger pg mumber.
>>
>> On Feb 3, 2016 8:09 PM, "M Ranga Swami Reddy" <swamireddy@xxxxxxxxx> wrote:
>>>
>>> Hi,
>>>
>>> I am using ceph for my storage cluster and health shows as WARN state
>>> with too few pgs.
>>>
>>> ==
>>> health HEALTH_WARN pool volumes has too few pgs ==
>>>
>>> The volume pool has 4096 pgs
>>> --
>>> ceph osd pool get volumes pg_num
>>> pg_num: 4096
>>> ---
>>>
>>> and
>>> >ceph df
>>>     NAME                   ID     USED      %USED     MAX AVAIL
>>> OBJECTS
>>>     volumes                4      2830G      0.82          108T
>>> 763509
>>> --
>>>
>>> How do we fix this, without downtime?
>>>
>>> Thanks
>>> Swami
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@xxxxxxxxxxxxxx
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> PLEASE NOTE: The information contained in this electronic mail message is intended only for the use of the designated recipient(s) named above. If the reader of this message is not the intended recipient, you are hereby notified that you have received this message in error and that any review, dissemination, distribution, or copying of this message is strictly prohibited. If you have received this communication in error, please notify the sender by telephone or e-mail (as shown above) immediately and destroy any and all copies of this message in your possession (whether hard copies or electronically stored copies).
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux