It runs by default in dry-run mode, which IMHO opinion should be the default for operations like this. IIRC you add "-d -r" to make it actually apply the re-weighting. Cheers, On 8 June 2016 at 15:04, M Ranga Swami Reddy <swamireddy@xxxxxxxxx> wrote: > Blair - Thanks for the script...Btw, is this script has option for dry run? > > Thanks > Swami > > On Wed, Jun 8, 2016 at 6:35 AM, Blair Bethwaite > <blair.bethwaite@xxxxxxxxx> wrote: >> Swami, >> >> Try https://github.com/cernceph/ceph-scripts/blob/master/tools/crush-reweight-by-utilization.py, >> that'll work with Firefly and allow you to only tune down weight of a >> specific number of overfull OSDs. >> >> Cheers, >> >> On 7 June 2016 at 23:11, M Ranga Swami Reddy <swamireddy@xxxxxxxxx> wrote: >>> OK, understood... >>> To fix the nearfull warn, I am reducing the weight of a specific OSD, >>> which filled >85%.. >>> Is this work-around advisable? >>> >>> Thanks >>> Swami >>> >>> On Tue, Jun 7, 2016 at 6:37 PM, Sage Weil <sage@xxxxxxxxxxxx> wrote: >>>> On Tue, 7 Jun 2016, M Ranga Swami Reddy wrote: >>>>> Hi Sage, >>>>> >Jewel and the latest hammer point release have an improved >>>>> >reweight-by-utilization (ceph osd test-reweight-by-utilization ... to dry >>>>> > run) to correct this. >>>>> >>>>> Thank you....But not planning to upgrade the cluster soon. >>>>> So, in this case - are there any tunable options will help? like >>>>> "crush tunable optimal" or so? >>>>> OR any other configuration options change will help? >>>> >>>> Firefly also has reweight-by-utilization... it's just a bit less friendly >>>> than the newer versions. CRUSH tunables don't generally help here unless >>>> you have lots of OSDs that are down+out. >>>> >>>> Note that firefly is no longer supported. >>>> >>>> sage >>>> >>>> >>>>> >>>>> >>>>> Thanks >>>>> Swami >>>>> >>>>> >>>>> On Tue, Jun 7, 2016 at 6:00 PM, Sage Weil <sage@xxxxxxxxxxxx> wrote: >>>>> > On Tue, 7 Jun 2016, M Ranga Swami Reddy wrote: >>>>> >> Hello, >>>>> >> I have aorund 100 OSDs in my ceph cluster. In this a few OSDs filled >>>>> >> with >85% of data and few OSDs filled with ~60%-70% of data. >>>>> >> >>>>> >> Any reason why the unevenly OSDs filling happned? do I need to any >>>>> >> tweaks on configuration to fix the above? Please advise. >>>>> >> >>>>> >> PS: Ceph version is - 0.80.7 >>>>> > >>>>> > Jewel and the latest hammer point release have an improved >>>>> > reweight-by-utilization (ceph osd test-reweight-by-utilization ... to dry >>>>> > run) to correct this. >>>>> > >>>>> > sage >>>>> > >>>>> -- >>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in >>>>> the body of a message to majordomo@xxxxxxxxxxxxxxx >>>>> More majordomo info at http://vger.kernel.org/majordomo-info.html >>>>> >>>>> >>> -- >>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in >>> the body of a message to majordomo@xxxxxxxxxxxxxxx >>> More majordomo info at http://vger.kernel.org/majordomo-info.html >> >> >> >> -- >> Cheers, >> ~Blairo -- Cheers, ~Blairo -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html