Hi Swami, That's a known issue, which I believe is much improved in Jewel thanks to a priority queue added somewhere in the OSD op path (I think). If I were you I'd be planning to get off Firefly and upgrade. Cheers, On 10 June 2016 at 12:08, M Ranga Swami Reddy <swamireddy@xxxxxxxxx> wrote: > Blair - Thanks for the details. I used to set the low priority for > recovery during the rebalance/recovery activity. > Even though I set the recovery_priority as 5 (instead of 1) and > client-op_priority set as 63, some of my customers complained that > their VMs are not reachable for a few mins/secs during the reblancing > task. Not sure, these low priority configurations are doing the job as > its. > > Thanks > Swami > > On Thu, Jun 9, 2016 at 5:50 PM, Blair Bethwaite > <blair.bethwaite@xxxxxxxxx> wrote: >> Swami, >> >> Run it with the help option for more context: >> "./crush-reweight-by-utilization.py --help". In your example below >> it's reporting to you what changes it would make to your OSD reweight >> values based on the default option settings (because you didn't >> specify any options). To make the script actually apply those weight >> changes you need the "-d -r" or "--doit --really" flags. >> >> If you want to get an idea of the impact that the weight changes will >> have before actually starting to move data then I suggest setting >> norecover and nobackfill (ceph osd set ...) on your cluster before >> making the weight changes, you can then examine "ceph -s" output >> (looking at "objects misplaced" to determine the scale of recovery >> required. Unset the flags once ready to start or back-out the reweight >> settings if you change your mind. You'll also want to lower these >> recovery and backfill tunables to reduce impact to client I/O (and if >> possible do not do this reweight change during peak I/O hours): >> ceph tell osd.* injectargs '--osd-max-backfills 1' >> ceph tell osd.* injectargs '--osd-max-recovery-threads 1' >> ceph tell osd.* injectargs '--osd-recovery-op-priority 1' >> ceph tell osd.* injectargs '--osd-client-op-priority 63' >> ceph tell osd.* injectargs '--osd-recovery-max-active 1' >> >> Cheers, >> >> On 9 June 2016 at 20:20, M Ranga Swami Reddy <swamireddy@xxxxxxxxx> wrote: >>> Hi Blari, >>> I ran the script and results are below: >>> == >>> ./crush-reweight-by-utilization.py >>> average_util: 0.587024, overload_util: 0.704429, underload_util: 0.587024. >>> reweighted: >>> 43 (0.852690 >= 0.704429) [1.000000 -> 0.950000] >>> 238 (0.845154 >= 0.704429) [1.000000 -> 0.950000] >>> 104 (0.827908 >= 0.704429) [1.000000 -> 0.950000] >>> 173 (0.817063 >= 0.704429) [1.000000 -> 0.950000] >>> == >>> >>> is the above scripts says to reweight 43 -> 0.95? >>> >>> Thanks >>> Swami >>> >>> On Wed, Jun 8, 2016 at 10:34 AM, M Ranga Swami Reddy >>> <swamireddy@xxxxxxxxx> wrote: >>>> Blair - Thanks for the script...Btw, is this script has option for dry run? >>>> >>>> Thanks >>>> Swami >>>> >>>> On Wed, Jun 8, 2016 at 6:35 AM, Blair Bethwaite >>>> <blair.bethwaite@xxxxxxxxx> wrote: >>>>> Swami, >>>>> >>>>> Try https://github.com/cernceph/ceph-scripts/blob/master/tools/crush-reweight-by-utilization.py, >>>>> that'll work with Firefly and allow you to only tune down weight of a >>>>> specific number of overfull OSDs. >>>>> >>>>> Cheers, >>>>> >>>>> On 7 June 2016 at 23:11, M Ranga Swami Reddy <swamireddy@xxxxxxxxx> wrote: >>>>>> OK, understood... >>>>>> To fix the nearfull warn, I am reducing the weight of a specific OSD, >>>>>> which filled >85%.. >>>>>> Is this work-around advisable? >>>>>> >>>>>> Thanks >>>>>> Swami >>>>>> >>>>>> On Tue, Jun 7, 2016 at 6:37 PM, Sage Weil <sage@xxxxxxxxxxxx> wrote: >>>>>>> On Tue, 7 Jun 2016, M Ranga Swami Reddy wrote: >>>>>>>> Hi Sage, >>>>>>>> >Jewel and the latest hammer point release have an improved >>>>>>>> >reweight-by-utilization (ceph osd test-reweight-by-utilization ... to dry >>>>>>>> > run) to correct this. >>>>>>>> >>>>>>>> Thank you....But not planning to upgrade the cluster soon. >>>>>>>> So, in this case - are there any tunable options will help? like >>>>>>>> "crush tunable optimal" or so? >>>>>>>> OR any other configuration options change will help? >>>>>>> >>>>>>> Firefly also has reweight-by-utilization... it's just a bit less friendly >>>>>>> than the newer versions. CRUSH tunables don't generally help here unless >>>>>>> you have lots of OSDs that are down+out. >>>>>>> >>>>>>> Note that firefly is no longer supported. >>>>>>> >>>>>>> sage >>>>>>> >>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> Thanks >>>>>>>> Swami >>>>>>>> >>>>>>>> >>>>>>>> On Tue, Jun 7, 2016 at 6:00 PM, Sage Weil <sage@xxxxxxxxxxxx> wrote: >>>>>>>> > On Tue, 7 Jun 2016, M Ranga Swami Reddy wrote: >>>>>>>> >> Hello, >>>>>>>> >> I have aorund 100 OSDs in my ceph cluster. In this a few OSDs filled >>>>>>>> >> with >85% of data and few OSDs filled with ~60%-70% of data. >>>>>>>> >> >>>>>>>> >> Any reason why the unevenly OSDs filling happned? do I need to any >>>>>>>> >> tweaks on configuration to fix the above? Please advise. >>>>>>>> >> >>>>>>>> >> PS: Ceph version is - 0.80.7 >>>>>>>> > >>>>>>>> > Jewel and the latest hammer point release have an improved >>>>>>>> > reweight-by-utilization (ceph osd test-reweight-by-utilization ... to dry >>>>>>>> > run) to correct this. >>>>>>>> > >>>>>>>> > sage >>>>>>>> > >>>>>>>> -- >>>>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in >>>>>>>> the body of a message to majordomo@xxxxxxxxxxxxxxx >>>>>>>> More majordomo info at http://vger.kernel.org/majordomo-info.html >>>>>>>> >>>>>>>> >>>>>> -- >>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in >>>>>> the body of a message to majordomo@xxxxxxxxxxxxxxx >>>>>> More majordomo info at http://vger.kernel.org/majordomo-info.html >>>>> >>>>> >>>>> >>>>> -- >>>>> Cheers, >>>>> ~Blairo >> >> >> >> -- >> Cheers, >> ~Blairo -- Cheers, ~Blairo _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com