Thats great.. Will try this.. Thanks Swami On Wed, Jun 8, 2016 at 10:38 AM, Blair Bethwaite <blair.bethwaite@xxxxxxxxx> wrote: > It runs by default in dry-run mode, which IMHO opinion should be the > default for operations like this. IIRC you add "-d -r" to make it > actually apply the re-weighting. > > Cheers, > > On 8 June 2016 at 15:04, M Ranga Swami Reddy <swamireddy@xxxxxxxxx> wrote: >> Blair - Thanks for the script...Btw, is this script has option for dry run? >> >> Thanks >> Swami >> >> On Wed, Jun 8, 2016 at 6:35 AM, Blair Bethwaite >> <blair.bethwaite@xxxxxxxxx> wrote: >>> Swami, >>> >>> Try https://github.com/cernceph/ceph-scripts/blob/master/tools/crush-reweight-by-utilization.py, >>> that'll work with Firefly and allow you to only tune down weight of a >>> specific number of overfull OSDs. >>> >>> Cheers, >>> >>> On 7 June 2016 at 23:11, M Ranga Swami Reddy <swamireddy@xxxxxxxxx> wrote: >>>> OK, understood... >>>> To fix the nearfull warn, I am reducing the weight of a specific OSD, >>>> which filled >85%.. >>>> Is this work-around advisable? >>>> >>>> Thanks >>>> Swami >>>> >>>> On Tue, Jun 7, 2016 at 6:37 PM, Sage Weil <sage@xxxxxxxxxxxx> wrote: >>>>> On Tue, 7 Jun 2016, M Ranga Swami Reddy wrote: >>>>>> Hi Sage, >>>>>> >Jewel and the latest hammer point release have an improved >>>>>> >reweight-by-utilization (ceph osd test-reweight-by-utilization ... to dry >>>>>> > run) to correct this. >>>>>> >>>>>> Thank you....But not planning to upgrade the cluster soon. >>>>>> So, in this case - are there any tunable options will help? like >>>>>> "crush tunable optimal" or so? >>>>>> OR any other configuration options change will help? >>>>> >>>>> Firefly also has reweight-by-utilization... it's just a bit less friendly >>>>> than the newer versions. CRUSH tunables don't generally help here unless >>>>> you have lots of OSDs that are down+out. >>>>> >>>>> Note that firefly is no longer supported. >>>>> >>>>> sage >>>>> >>>>> >>>>>> >>>>>> >>>>>> Thanks >>>>>> Swami >>>>>> >>>>>> >>>>>> On Tue, Jun 7, 2016 at 6:00 PM, Sage Weil <sage@xxxxxxxxxxxx> wrote: >>>>>> > On Tue, 7 Jun 2016, M Ranga Swami Reddy wrote: >>>>>> >> Hello, >>>>>> >> I have aorund 100 OSDs in my ceph cluster. In this a few OSDs filled >>>>>> >> with >85% of data and few OSDs filled with ~60%-70% of data. >>>>>> >> >>>>>> >> Any reason why the unevenly OSDs filling happned? do I need to any >>>>>> >> tweaks on configuration to fix the above? Please advise. >>>>>> >> >>>>>> >> PS: Ceph version is - 0.80.7 >>>>>> > >>>>>> > Jewel and the latest hammer point release have an improved >>>>>> > reweight-by-utilization (ceph osd test-reweight-by-utilization ... to dry >>>>>> > run) to correct this. >>>>>> > >>>>>> > sage >>>>>> > >>>>>> -- >>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in >>>>>> the body of a message to majordomo@xxxxxxxxxxxxxxx >>>>>> More majordomo info at http://vger.kernel.org/majordomo-info.html >>>>>> >>>>>> >>>> -- >>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in >>>> the body of a message to majordomo@xxxxxxxxxxxxxxx >>>> More majordomo info at http://vger.kernel.org/majordomo-info.html >>> >>> >>> >>> -- >>> Cheers, >>> ~Blairo > > > > -- > Cheers, > ~Blairo _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com