Re: ceph mon_data_size_warn limits for large cluster

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On older releases, at least, inflated DBs correlated with miserable recovery performance and lots of slow requests.  The  DB and OSDs were also on HDD FWIW.   A single drive failure would result in substantial RBD impact.  

> On Feb 18, 2019, at 3:28 AM, Dan van der Ster <dan@xxxxxxxxxxxxxx> wrote:
> 
> Not really.
> 
> You should just restart your mons though -- if done one at a time it
> has zero impact on your clients.
> 
> -- dan
> 
> 
> On Mon, Feb 18, 2019 at 12:11 PM M Ranga Swami Reddy
> <swamireddy@xxxxxxxxx> wrote:
>> 
>> Hi Sage - If the mon data increases, is this impacts the ceph cluster
>> performance (ie on ceph osd bench, etc)?
>> 
>> On Fri, Feb 15, 2019 at 3:13 PM M Ranga Swami Reddy
>> <swamireddy@xxxxxxxxx> wrote:
>>> 
>>> today I again hit the warn with 30G also...
>>> 
>>>> On Thu, Feb 14, 2019 at 7:39 PM Sage Weil <sage@xxxxxxxxxxxx> wrote:
>>>> 
>>>>> On Thu, 7 Feb 2019, Dan van der Ster wrote:
>>>>> On Thu, Feb 7, 2019 at 12:17 PM M Ranga Swami Reddy
>>>>> <swamireddy@xxxxxxxxx> wrote:
>>>>>> 
>>>>>> Hi Dan,
>>>>>>> During backfilling scenarios, the mons keep old maps and grow quite
>>>>>>> quickly. So if you have balancing, pg splitting, etc. ongoing for
>>>>>>> awhile, the mon stores will eventually trigger that 15GB alarm.
>>>>>>> But the intended behavior is that once the PGs are all active+clean,
>>>>>>> the old maps should be trimmed and the disk space freed.
>>>>>> 
>>>>>> old maps not trimmed after cluster reached to "all+clean" state for all PGs.
>>>>>> Is there (known) bug here?
>>>>>> As the size of dB showing > 15G, do I need to run the compact commands
>>>>>> to do the trimming?
>>>>> 
>>>>> Compaction isn't necessary -- you should only need to restart all
>>>>> peon's then the leader. A few minutes later the db's should start
>>>>> trimming.
>>>> 
>>>> The next time someone sees this behavior, can you please
>>>> 
>>>> - enable debug_mon = 20 on all mons (*before* restarting)
>>>>   ceph tell mon.* injectargs '--debug-mon 20'
>>>> - wait for 10 minutes or so to generate some logs
>>>> - add 'debug mon = 20' to ceph.conf (on mons only)
>>>> - restart the monitors
>>>> - wait for them to start trimming
>>>> - remove 'debug mon = 20' from ceph.conf (on mons only)
>>>> - tar up the log files, ceph-post-file them, and share them with ticket
>>>> http://tracker.ceph.com/issues/38322
>>>> 
>>>> Thanks!
>>>> sage
>>>> 
>>>> 
>>>> 
>>>> 
>>>>> -- dan
>>>>> 
>>>>> 
>>>>>> 
>>>>>> Thanks
>>>>>> Swami
>>>>>> 
>>>>>>> On Wed, Feb 6, 2019 at 6:24 PM Dan van der Ster <dan@xxxxxxxxxxxxxx> wrote:
>>>>>>> 
>>>>>>> Hi,
>>>>>>> 
>>>>>>> With HEALTH_OK a mon data dir should be under 2GB for even such a large cluster.
>>>>>>> 
>>>>>>> During backfilling scenarios, the mons keep old maps and grow quite
>>>>>>> quickly. So if you have balancing, pg splitting, etc. ongoing for
>>>>>>> awhile, the mon stores will eventually trigger that 15GB alarm.
>>>>>>> But the intended behavior is that once the PGs are all active+clean,
>>>>>>> the old maps should be trimmed and the disk space freed.
>>>>>>> 
>>>>>>> However, several people have noted that (at least in luminous
>>>>>>> releases) the old maps are not trimmed until after HEALTH_OK *and* all
>>>>>>> mons are restarted. This ticket seems related:
>>>>>>> http://tracker.ceph.com/issues/37875
>>>>>>> 
>>>>>>> (Over here we're restarting mons every ~2-3 weeks, resulting in the
>>>>>>> mon stores dropping from >15GB to ~700MB each time).
>>>>>>> 
>>>>>>> -- Dan
>>>>>>> 
>>>>>>> 
>>>>>>>> On Wed, Feb 6, 2019 at 1:26 PM Sage Weil <sage@xxxxxxxxxxxx> wrote:
>>>>>>>> 
>>>>>>>> Hi Swami
>>>>>>>> 
>>>>>>>> The limit is somewhat arbitrary, based on cluster sizes we had seen when
>>>>>>>> we picked it.  In your case it should be perfectly safe to increase it.
>>>>>>>> 
>>>>>>>> sage
>>>>>>>> 
>>>>>>>> 
>>>>>>>>> On Wed, 6 Feb 2019, M Ranga Swami Reddy wrote:
>>>>>>>>> 
>>>>>>>>> Hello -  Are the any limits for mon_data_size for cluster with 2PB
>>>>>>>>> (with 2000+ OSDs)?
>>>>>>>>> 
>>>>>>>>> Currently it set as 15G. What is logic behind this? Can we increase
>>>>>>>>> when we get the mon_data_size_warn messages?
>>>>>>>>> 
>>>>>>>>> I am getting the mon_data_size_warn message even though there a ample
>>>>>>>>> of free space on the disk (around 300G free disk)
>>>>>>>>> 
>>>>>>>>> Earlier thread on the same discusion:
>>>>>>>>> https://www.spinics.net/lists/ceph-users/msg42456.html
>>>>>>>>> 
>>>>>>>>> Thanks
>>>>>>>>> Swami
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> _______________________________________________
>>>>>>>> ceph-users mailing list
>>>>>>>> ceph-users@xxxxxxxxxxxxxx
>>>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>> 
>>>>> 
>>>>> 
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux