Re: "store is getting too big" on monitors

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 02/16/2015 12:57 PM, Mohamed Pakkeer wrote:

  Hi ceph-experts,

   We are getting "store is getting too big" on our test cluster.
Cluster is running with giant release and configured as EC pool to test
cephFS.

cluster c2a97a2f-fdc7-4eb5-82ef-70c52f2eceb1
      health HEALTH_WARN too few pgs per osd (0 < min 20); mon.master01
store is getting too big! 15376 MB >= 15360 MB; mon.master02 store is
getting too big! 15402 MB >= 15360 MB; mon.master03 store is getting too
big! 15402 MB >= 15360 MB; clock skew detected on mon.master02, mon.master03
      monmap e3: 3 mons at
{master01=10.1.2.231:6789/0,master02=10.1.2.232:6789/0,master03=10.1.2.233:6789/0
<http://10.1.2.231:6789/0,master02=10.1.2.232:6789/0,master03=10.1.2.233:6789/0>},
election epoch 38, quorum 0,1,2 master01,master02,master03
      osdmap e97396: 552 osds: 552 up, 552 in
       pgmap v354736: 0 pgs, 0 pools, 0 bytes data, 0 objects
             8547 GB used, 1953 TB / 1962 TB avail

We tried monitor restart with mon compact on start = true as well as
manual compaction using 'ceph tell mon.FOO compact'. But it didn't
reduce the size of store.db. We already deleted the pools and mds to
start fresh cluster. Do we need to delete the mon and recreate again or
do we have any solution to reduce the store size?

Could you get us a list of all the keys on the store using 'ceph-kvstore-tool' ? Instructions on the email you quoted.

Cheers!

  -Joao


Regards,
K.Mohamed Pakkeer



On 12/10/2014 07:30 PM, Kevin Sumner wrote:

    The mons have grown another 30GB each overnight (except for 003?), which
    is quite worrying.  I ran a little bit of testing yesterday after my
    post, but not a significant amount.

    I wouldn’t expect compact on start to help this situation based on the
    name since we don’t (shouldn’t?) restart the mons regularly, but there
    appears to be no documentation on it.  We’re pretty good on disk space
    on the mons currently, but if that changes, I’ll probably use this to
    see about bringing these numbers in line.

This is an issue that has been seen on larger clusters, and it usually
takes a monitor restart, with 'mon compact on start = true' or manual
compaction 'ceph tell mon.FOO compact' to bring the monitor back to a
sane disk usage level.

However, I have not been able to reproduce this in order to track the
source. I'm guessing I lack the scale of the cluster, or the appropriate
workload (maybe both).

What kind of workload are you running the cluster through? You mention
cephfs, but do you have any more info you can share that could help us
reproducing this state?

Sage also fixed an issue that could potentially cause this (depending on
what is causing it in the first place) [1,2,3]. This bug, #9987, is due
to a given cached value not being updated, leading to the monitor not
removing unnecessary data, potentially causing this growth. This cached
value would be set to its proper value when the monitor is restarted
though, so a simple restart would have all this unnecessary data blown away.

Restarting the monitor ends up masking the true cause of the store
growth: whether from #9987 or from obsolete data kept by the monitor's
backing store (leveldb), either due to misuse of leveldb or due to
leveldb's nature (haven't been able to ascertain which may be at fault,
partly due to being unable to reproduce the problem).

If you are up to it, I would suggest the following approach in hope to
determine what may be at fault:

1) 'ceph tell mon.FOO compact' -- which will force the monitor to
compact its store. This won't close leveldb, so it won't have much
effect on the store size if it happens to be leveldb holding on to some
data (I could go into further detail, but I don't think this is the
right medium). 1.a) you may notice the store increasing in size during
this period; it's expected. 1.b) compaction may take a while, but in the
end you'll hopefully see a significant reduction in size.

2) Assuming that failed, I would suggest doing the following:

2.1) grab ceph-kvstore-tool from the ceph-test package
2.2) stop the monitor

2.3) run 'ceph-kvstore-tool /var/lib/ceph/mon/ceph-FOO/store.db list >
store.dump'

2.4) run (for above store's location, let's call it $STORE:

for m in osdmap pgmap; do
   for k in first_committed last_committed; do
     ceph-kvstore-tool $STORE get $m $k >> store.dump
   done
done

ceph-kvstore-tool $STORE get pgmap_meta last_osdmap_epoch >> store.dump
ceph-kvstore-tool $STORE get pgmap_meta version >> store.dump

2.5) send over the results of the dump

2.6) if you were to compress the store as well and send me a link to
grab it I would appreciate it.

3) Next you could simply restart the monitor (without 'mon compact on
start = true'); if the monitor's store size decreases, then there's a
fair chance that you've been bit by #9987. Otherwise, it may be
leveldb's clutter. You should also note that leveldb may itself compact
automatically on start, so it's hard to say for sure what fixed what.

4) If store size hasn't gone back to sane levels by now, you may wish to
restart with 'mon compact on start = true' and see if it helps. If it
doesn't, then we may have a completely different issue in our hands.

Now, assuming your store size went down on step 3, and if you are
willing, it would be interesting to see if Sage's patches helps out in
any way. The patches have not been backported to the giant branch yet,
so you would have to apply them yourself. For them to work you would
have to run the patched monitor as the leader. I would suggest leaving
the other monitors running an unpatched version so they could act as the
control group.

Let us know if any of this helps.

Cheers!

   -Joao

[1] -http://tracker.ceph.com/issues/9987
[2] - 093c5f0cabeb552b90d944da2c50de48fcf6f564
[3] - 3fb731b722c50672a5a9de0c86a621f5f50f2d06

    :: ~ » ceph health detail | grep 'too big'
    HEALTH_WARN mon.cluster4-monitor001 store is getting too big! 77365 MB
      >= 15360 MB; mon.cluster4-monitor002 store is getting too big! 87868 MB
      >= 15360 MB; mon.cluster4-monitor003 store is getting too big! 30359 MB
      >= 15360 MB; mon.cluster4-monitor004 store is getting too big! 93414 MB
      >= 15360 MB; mon.cluster4-monitor005 store is getting too big! 88232 MB
      >= 15360 MB
    mon.cluster4-monitor001 store is getting too big! 77365 MB >= 15360 MB
    -- 72% avail
    mon.cluster4-monitor002 store is getting too big! 87868 MB >= 15360 MB
    -- 70% avail
    mon.cluster4-monitor003 store is getting too big! 30359 MB >= 15360 MB
    -- 85% avail
    mon.cluster4-monitor004 store is getting too big! 93414 MB >= 15360 MB
    -- 69% avail
    mon.cluster4-monitor005 store is getting too big! 88232 MB >= 15360 MB
    -- 71% avail
    --
    Kevin Sumner
    ke...@xxxxxxxxx  <mailto:ke...@xxxxxxxxx>  <mailto:ke...@xxxxxxxxx>



        On Dec 9, 2014, at 6:20 PM, Haomai Wang <haomaiw...@xxxxxxxxx  <mailto:haomaiw...@xxxxxxxxx>
        <mailto:haomaiw...@xxxxxxxxx>> wrote:

        Maybe you can enable "mon_compact_on_start=true" when restarting mon,
        it will compact data

        On Wed, Dec 10, 2014 at 6:50 AM, Kevin Sumner <ke...@xxxxxxxxx  <mailto:ke...@xxxxxxxxx>
        <mailto:ke...@xxxxxxxxx>> wrote:

            Hi all,

            We recently upgraded our cluster to Giant from.  Since then, we’ve been
            driving load tests against CephFS.  However, we’re getting “store is
            getting
            too big” warnings from the monitors and the mons have started
            consuming way
            more disk space, 40GB-60GB now as opposed to ~10GB pre-upgrade.  Is this
            expected?  Is there anything I can do to ease the store’s size?

            Thanks!

            :: ~ » ceph status
                cluster f1aefa73-b968-41e0-9a28-9a465db5f10b
                 health HEALTH_WARN mon.cluster4-monitor001 store is getting too big!
            45648 MB >= 15360 MB; mon.cluster4-monitor002 store is getting too big!
            56939 MB >= 15360 MB; mon.cluster4-monitor003 store is getting too big!
            28647 MB >= 15360 MB; mon.cluster4-monitor004 store is getting too big!
            60655 MB >= 15360 MB; mon.cluster4-monitor005 store is getting too big!
            57335 MB >= 15360 MB
                 monmap e3: 5 mons at
            {cluster4-monitor001=17.138.96.12:6789/0,cluster4-monitor002=17.138.96.13:6789/0,cluster4-monitor003=17.138.96.14:6789/0,cluster4-monitor004=17.138.96.15:6789/0,cluster4-monitor005=17.138.96.16:6789/0  <http://17.138.96.12:6789/0,cluster4-monitor002=17.138.96.13:6789/0,cluster4-monitor003=17.138.96.14:6789/0,cluster4-monitor004=17.138.96.15:6789/0,cluster4-monitor005=17.138.96.16:6789/0>},
            election epoch 34938, quorum 0,1,2,3,4
            cluster4-monitor001,cluster4-monitor002,cluster4-monitor003,cluster4-monitor004,cluster4-monitor005
                 mdsmap e6538: 1/1/1 up {0=cluster4-monitor001=up:active}
                 osdmap e49500: 501 osds: 470 up, 469 in
                  pgmap v1369307: 98304 pgs, 3 pools, 4933 GB data, 1976 kobjects
                        16275 GB used, 72337 GB / 93366 GB avail
                           98304 active+clean
              client io 3463 MB/s rd, 18710 kB/s wr, 7456 op/s
            --
            Kevin Sumner
            ke...@xxxxxxxxx  <mailto:ke...@xxxxxxxxx>  <mailto:ke...@xxxxxxxxx>




            _______________________________________________
            ceph-users mailing list
            ceph-users@xxxxxxxxxxxxxx  <mailto:ceph-users@xxxxxxxxxxxxxx>
            http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

        --
        Best Regards,

        Wheat

    _______________________________________________
    ceph-users mailing list
    ceph-users@xxxxxxxxxxxxxx  <mailto:ceph-users@xxxxxxxxxxxxxx>
    http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx  <mailto:ceph-users@xxxxxxxxxxxxxx>
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


--
Thanks & Regards
K.Mohamed Pakkeer
Mobile- 0091-8754410114


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com





[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux