Re: 6 pgs not deep-scrubbed in time

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



And as said before still it is in warning state with pgs not deep-scrubed
in time . Hope this can be ignored and set those two flags "noout and
nobackfill" then reboot .

Thank you again Sir

On Thu, 1 Feb 2024, 16:11 Michel Niyoyita, <micou12@xxxxxxxxx> wrote:

> Thank you very much Janne.
>
> On Thu, 1 Feb 2024, 15:21 Janne Johansson, <icepic.dz@xxxxxxxxx> wrote:
>
>> pause and nodown is not a good option to set, that will certainly make
>> clients stop IO. Pause will stop it immediately, and nodown will stop
>> IO when the OSD processes stop running on this host.
>>
>> When we do service on a host, we set "noout" and "nobackfill", that is
>> enough for reboots, OS upgrades and simple disk exchanges.
>> The PGs on this one host will be degraded during the down period, but
>> IO continues.
>> Of course this is when the cluster was healthy to begin with (not
>> counting "not scrubbed in time" warnings, they don't matter in this
>> case.)
>>
>>
>>
>> Den tors 1 feb. 2024 kl 12:21 skrev Michel Niyoyita <micou12@xxxxxxxxx>:
>> >
>> > Thanks Very much Wesley,
>> >
>> > We have decided to restart one host among three osds hosts. before doing
>> > that I need the advices of the team . these are flags I want to set
>> before
>> > restart.
>> >
>> >  'ceph osd set noout'
>> >  'ceph osd set nobackfill'
>> >  'ceph osd set norecover'
>> >  'ceph osd set norebalance'
>> > 'ceph osd set nodown'
>> >  'ceph osd set pause'
>> > 'ceph osd set nodeep-scrub'
>> > 'ceph osd set noscrub'
>> >
>> >
>> > Would like to ask if this can be enough to set and restart the host
>> safely
>> > . the cluster has 3 as replicas.
>> >
>> > will the cluster still be accessible while restart the hosts? after
>> > restarting I will unset the flags.
>> >
>> > Kindly advise.
>> >
>> > Michel
>> >
>> >
>> > On Tue, 30 Jan 2024, 17:44 Wesley Dillingham, <wes@xxxxxxxxxxxxxxxxx>
>> wrote:
>> >
>> > > actually it seems the issue I had in mind was fixed in 16.2.11 so you
>> > > should be fine.
>> > >
>> > > Respectfully,
>> > >
>> > > *Wes Dillingham*
>> > > wes@xxxxxxxxxxxxxxxxx
>> > > LinkedIn <http://www.linkedin.com/in/wesleydillingham>
>> > >
>> > >
>> > > On Tue, Jan 30, 2024 at 10:34 AM Wesley Dillingham <
>> wes@xxxxxxxxxxxxxxxxx>
>> > > wrote:
>> > >
>> > >> You may want to consider upgrading to 16.2.14 before you do the pg
>> split.
>> > >>
>> > >> Respectfully,
>> > >>
>> > >> *Wes Dillingham*
>> > >> wes@xxxxxxxxxxxxxxxxx
>> > >> LinkedIn <http://www.linkedin.com/in/wesleydillingham>
>> > >>
>> > >>
>> > >> On Tue, Jan 30, 2024 at 10:18 AM Michel Niyoyita <micou12@xxxxxxxxx>
>> > >> wrote:
>> > >>
>> > >>> I tried that on one of my pool (pool id 3) but the number of pgs not
>> > >>> deep-scrubbed in time increased also from 55 to 100 but the number
>> of PGs
>> > >>> was increased. I set also autoscale to off mode. before continue to
>> other
>> > >>> pools would like to ask if so far there is no negative impact.
>> > >>>
>> > >>> ceph -s
>> > >>>   cluster:
>> > >>>     id:     cb0caedc-eb5b-42d1-a34f-96facfda8c27
>> > >>>     health: HEALTH_WARN
>> > >>>             100 pgs not deep-scrubbed in time
>> > >>>
>> > >>>   services:
>> > >>>     mon: 3 daemons, quorum ceph-mon1,ceph-mon2,ceph-mon3 (age 11M)
>> > >>>     mgr: ceph-mon2(active, since 11M), standbys: ceph-mon3,
>> ceph-mon1
>> > >>>     osd: 48 osds: 48 up (since 11M), 48 in (since 12M)
>> > >>>     rgw: 6 daemons active (6 hosts, 1 zones)
>> > >>>
>> > >>>   data:
>> > >>>     pools:   10 pools, 609 pgs
>> > >>>     objects: 6.03M objects, 23 TiB
>> > >>>     usage:   151 TiB used, 282 TiB / 433 TiB avail
>> > >>>     pgs:     603 active+clean
>> > >>>              4   active+clean+scrubbing+deep
>> > >>>              2   active+clean+scrubbing
>> > >>>
>> > >>>   io:
>> > >>>     client:   96 MiB/s rd, 573 MiB/s wr, 576 op/s rd, 648 op/s wr
>> > >>>
>> > >>> root@ceph-osd3:/var/log# ceph df
>> > >>> --- RAW STORAGE ---
>> > >>> CLASS     SIZE    AVAIL     USED  RAW USED  %RAW USED
>> > >>> hdd    433 TiB  282 TiB  151 TiB   151 TiB      34.93
>> > >>> TOTAL  433 TiB  282 TiB  151 TiB   151 TiB      34.93
>> > >>>
>> > >>> --- POOLS ---
>> > >>> POOL                   ID  PGS   STORED  OBJECTS     USED  %USED
>> MAX
>> > >>> AVAIL
>> > >>> device_health_metrics   1    1  1.1 MiB        3  3.2 MiB      0
>>  72
>> > >>> TiB
>> > >>> .rgw.root               2   32  3.7 KiB        8   96 KiB      0
>>  72
>> > >>> TiB
>> > >>> default.rgw.log         3  256  3.6 KiB      204  408 KiB      0
>>  72
>> > >>> TiB
>> > >>> default.rgw.control     4   32      0 B        8      0 B      0
>>  72
>> > >>> TiB
>> > >>> default.rgw.meta        5   32    382 B        2   24 KiB      0
>>  72
>> > >>> TiB
>> > >>> volumes                 6  128   21 TiB    5.74M   62 TiB  22.30
>>  72
>> > >>> TiB
>> > >>> images                  7   32  878 GiB  112.50k  2.6 TiB   1.17
>>  72
>> > >>> TiB
>> > >>> backups                 8   32      0 B        0      0 B      0
>>  72
>> > >>> TiB
>> > >>> vms                     9   32  870 GiB  170.73k  2.5 TiB   1.13
>>  72
>> > >>> TiB
>> > >>> testbench              10   32      0 B        0      0 B      0
>>  72
>> > >>> TiB
>> > >>>
>> > >>> On Tue, Jan 30, 2024 at 5:05 PM Wesley Dillingham <
>> wes@xxxxxxxxxxxxxxxxx>
>> > >>> wrote:
>> > >>>
>> > >>>> It will take a couple weeks to a couple months to complete is my
>> best
>> > >>>> guess on 10TB spinners at ~40% full. The cluster should be usable
>> > >>>> throughout the process.
>> > >>>>
>> > >>>> Keep in mind, you should disable the pg autoscaler on any pool
>> which
>> > >>>> you are manually adjusting the pg_num for. Increasing the pg_num
>> is called
>> > >>>> "pg splitting" you can google around for this to see how it will
>> work etc.
>> > >>>>
>> > >>>> There are a few knobs to increase or decrease the aggressiveness
>> of the
>> > >>>> pg split, primarily these are osd_max_backfills and
>> > >>>> target_max_misplaced_ratio.
>> > >>>>
>> > >>>> You can monitor the progress of the split by looking at "ceph osd
>> pool
>> > >>>> ls detail" for the pool you are splitting, for this pool pgp_num
>> will
>> > >>>> slowly increase up until it reaches the pg_num / pg_num_target.
>> > >>>>
>> > >>>> IMO this blog post best covers the issue which you are looking to
>> > >>>> undertake:
>> > >>>>
>> https://ceph.io/en/news/blog/2019/new-in-nautilus-pg-merging-and-autotuning/
>> > >>>>
>> > >>>> Respectfully,
>> > >>>>
>> > >>>> *Wes Dillingham*
>> > >>>> wes@xxxxxxxxxxxxxxxxx
>> > >>>> LinkedIn <http://www.linkedin.com/in/wesleydillingham>
>> > >>>>
>> > >>>>
>> > >>>> On Tue, Jan 30, 2024 at 9:38 AM Michel Niyoyita <micou12@xxxxxxxxx
>> >
>> > >>>> wrote:
>> > >>>>
>> > >>>>> Thanks for your advices Wes, below is what ceph osd df tree shows
>> ,
>> > >>>>> the increase of pg_num of the production cluster will not affect
>> the
>> > >>>>> performance or crush ? how long it can takes to finish?
>> > >>>>>
>> > >>>>> ceph osd df tree
>> > >>>>> ID  CLASS  WEIGHT     REWEIGHT  SIZE     RAW USE  DATA      OMAP
>> > >>>>>  META     AVAIL    %USE   VAR   PGS  STATUS  TYPE NAME
>> > >>>>> -1         433.11841         -  433 TiB  151 TiB    67 TiB  364
>> MiB
>> > >>>>> 210 GiB  282 TiB  34.86  1.00    -          root default
>> > >>>>> -3         144.37280         -  144 TiB   50 TiB    22 TiB  121
>> MiB
>> > >>>>>  70 GiB   94 TiB  34.86  1.00    -              host ceph-osd1
>> > >>>>>  2    hdd    9.02330   1.00000  9.0 TiB  2.7 TiB  1021 GiB  5.4
>> MiB
>> > >>>>> 3.7 GiB  6.3 TiB  30.40  0.87   19      up          osd.2
>> > >>>>>  3    hdd    9.02330   1.00000  9.0 TiB  2.7 TiB   931 GiB  4.1
>> MiB
>> > >>>>> 3.5 GiB  6.4 TiB  29.43  0.84   29      up          osd.3
>> > >>>>>  6    hdd    9.02330   1.00000  9.0 TiB  3.3 TiB   1.5 TiB  8.1
>> MiB
>> > >>>>> 4.5 GiB  5.8 TiB  36.09  1.04   20      up          osd.6
>> > >>>>>  9    hdd    9.02330   1.00000  9.0 TiB  2.8 TiB   1.0 TiB  6.6
>> MiB
>> > >>>>> 3.8 GiB  6.2 TiB  30.97  0.89   23      up          osd.9
>> > >>>>> 12    hdd    9.02330   1.00000  9.0 TiB  4.0 TiB   2.3 TiB   13
>> MiB
>> > >>>>> 6.1 GiB  5.0 TiB  44.68  1.28   30      up          osd.12
>> > >>>>> 15    hdd    9.02330   1.00000  9.0 TiB  3.5 TiB   1.8 TiB  9.2
>> MiB
>> > >>>>> 5.2 GiB  5.5 TiB  38.99  1.12   30      up          osd.15
>> > >>>>> 18    hdd    9.02330   1.00000  9.0 TiB  3.0 TiB   1.2 TiB  6.5
>> MiB
>> > >>>>> 4.0 GiB  6.1 TiB  32.80  0.94   21      up          osd.18
>> > >>>>> 22    hdd    9.02330   1.00000  9.0 TiB  3.6 TiB   1.9 TiB   10
>> MiB
>> > >>>>> 5.4 GiB  5.4 TiB  40.25  1.15   22      up          osd.22
>> > >>>>> 25    hdd    9.02330   1.00000  9.0 TiB  3.9 TiB   2.1 TiB   12
>> MiB
>> > >>>>> 5.7 GiB  5.1 TiB  42.94  1.23   22      up          osd.25
>> > >>>>> 28    hdd    9.02330   1.00000  9.0 TiB  3.1 TiB   1.4 TiB  7.5
>> MiB
>> > >>>>> 4.1 GiB  5.9 TiB  34.87  1.00   21      up          osd.28
>> > >>>>> 32    hdd    9.02330   1.00000  9.0 TiB  2.7 TiB  1017 GiB  4.8
>> MiB
>> > >>>>> 3.7 GiB  6.3 TiB  30.36  0.87   27      up          osd.32
>> > >>>>> 35    hdd    9.02330   1.00000  9.0 TiB  3.0 TiB   1.3 TiB  7.2
>> MiB
>> > >>>>> 4.2 GiB  6.0 TiB  33.73  0.97   21      up          osd.35
>> > >>>>> 38    hdd    9.02330   1.00000  9.0 TiB  3.1 TiB   1.4 TiB  7.3
>> MiB
>> > >>>>> 4.1 GiB  5.9 TiB  34.57  0.99   24      up          osd.38
>> > >>>>> 41    hdd    9.02330   1.00000  9.0 TiB  2.9 TiB   1.2 TiB  6.2
>> MiB
>> > >>>>> 4.0 GiB  6.1 TiB  32.49  0.93   24      up          osd.41
>> > >>>>> 44    hdd    9.02330   1.00000  9.0 TiB  3.1 TiB   1.4 TiB  7.3
>> MiB
>> > >>>>> 4.4 GiB  5.9 TiB  34.87  1.00   29      up          osd.44
>> > >>>>> 47    hdd    9.02330   1.00000  9.0 TiB  2.7 TiB  1016 GiB  5.4
>> MiB
>> > >>>>> 3.6 GiB  6.3 TiB  30.35  0.87   23      up          osd.47
>> > >>>>> -7         144.37280         -  144 TiB   50 TiB    22 TiB  122
>> MiB
>> > >>>>>  70 GiB   94 TiB  34.86  1.00    -              host ceph-osd2
>> > >>>>>  1    hdd    9.02330   1.00000  9.0 TiB  2.8 TiB   1.1 TiB  5.7
>> MiB
>> > >>>>> 3.8 GiB  6.2 TiB  31.00  0.89   27      up          osd.1
>> > >>>>>  5    hdd    9.02330   1.00000  9.0 TiB  3.2 TiB   1.5 TiB  7.3
>> MiB
>> > >>>>> 4.5 GiB  5.8 TiB  35.45  1.02   27      up          osd.5
>> > >>>>>  8    hdd    9.02330   1.00000  9.0 TiB  3.3 TiB   1.6 TiB  8.3
>> MiB
>> > >>>>> 4.7 GiB  5.7 TiB  36.85  1.06   30      up          osd.8
>> > >>>>> 10    hdd    9.02330   1.00000  9.0 TiB  3.1 TiB   1.4 TiB  7.5
>> MiB
>> > >>>>> 4.5 GiB  5.9 TiB  34.87  1.00   20      up          osd.10
>> > >>>>> 13    hdd    9.02330   1.00000  9.0 TiB  3.6 TiB   1.8 TiB   10
>> MiB
>> > >>>>> 5.3 GiB  5.4 TiB  39.63  1.14   27      up          osd.13
>> > >>>>> 16    hdd    9.02330   1.00000  9.0 TiB  2.8 TiB   1.1 TiB  6.0
>> MiB
>> > >>>>> 3.8 GiB  6.2 TiB  31.01  0.89   19      up          osd.16
>> > >>>>> 19    hdd    9.02330   1.00000  9.0 TiB  3.0 TiB   1.2 TiB  6.4
>> MiB
>> > >>>>> 4.0 GiB  6.1 TiB  32.77  0.94   21      up          osd.19
>> > >>>>> 21    hdd    9.02330   1.00000  9.0 TiB  2.8 TiB   1.1 TiB  5.5
>> MiB
>> > >>>>> 3.7 GiB  6.2 TiB  31.58  0.91   26      up          osd.21
>> > >>>>> 24    hdd    9.02330   1.00000  9.0 TiB  2.6 TiB   855 GiB  4.7
>> MiB
>> > >>>>> 3.3 GiB  6.4 TiB  28.61  0.82   19      up          osd.24
>> > >>>>> 27    hdd    9.02330   1.00000  9.0 TiB  3.7 TiB   1.9 TiB   10
>> MiB
>> > >>>>> 5.2 GiB  5.3 TiB  40.84  1.17   24      up          osd.27
>> > >>>>> 30    hdd    9.02330   1.00000  9.0 TiB  3.2 TiB   1.4 TiB  7.5
>> MiB
>> > >>>>> 4.5 GiB  5.9 TiB  35.16  1.01   22      up          osd.30
>> > >>>>> 33    hdd    9.02330   1.00000  9.0 TiB  3.1 TiB   1.4 TiB  8.6
>> MiB
>> > >>>>> 4.3 GiB  5.9 TiB  34.59  0.99   23      up          osd.33
>> > >>>>> 36    hdd    9.02330   1.00000  9.0 TiB  3.4 TiB   1.7 TiB   10
>> MiB
>> > >>>>> 5.0 GiB  5.6 TiB  38.17  1.09   25      up          osd.36
>> > >>>>> 39    hdd    9.02330   1.00000  9.0 TiB  3.4 TiB   1.7 TiB  8.5
>> MiB
>> > >>>>> 5.1 GiB  5.6 TiB  37.79  1.08   31      up          osd.39
>> > >>>>> 42    hdd    9.02330   1.00000  9.0 TiB  3.6 TiB   1.8 TiB   10
>> MiB
>> > >>>>> 5.2 GiB  5.4 TiB  39.68  1.14   23      up          osd.42
>> > >>>>> 45    hdd    9.02330   1.00000  9.0 TiB  2.7 TiB   964 GiB  5.1
>> MiB
>> > >>>>> 3.5 GiB  6.3 TiB  29.78  0.85   21      up          osd.45
>> > >>>>> -5         144.37280         -  144 TiB   50 TiB    22 TiB  121
>> MiB
>> > >>>>>  70 GiB   94 TiB  34.86  1.00    -              host ceph-osd3
>> > >>>>>  0    hdd    9.02330   1.00000  9.0 TiB  2.7 TiB   934 GiB  4.9
>> MiB
>> > >>>>> 3.4 GiB  6.4 TiB  29.47  0.85   21      up          osd.0
>> > >>>>>  4    hdd    9.02330   1.00000  9.0 TiB  3.0 TiB   1.2 TiB  6.5
>> MiB
>> > >>>>> 4.1 GiB  6.1 TiB  32.73  0.94   22      up          osd.4
>> > >>>>>  7    hdd    9.02330   1.00000  9.0 TiB  3.5 TiB   1.8 TiB  9.2
>> MiB
>> > >>>>> 5.1 GiB  5.5 TiB  39.02  1.12   30      up          osd.7
>> > >>>>> 11    hdd    9.02330   1.00000  9.0 TiB  3.6 TiB   1.9 TiB   10
>> MiB
>> > >>>>> 5.1 GiB  5.4 TiB  39.97  1.15   27      up          osd.11
>> > >>>>> 14    hdd    9.02330   1.00000  9.0 TiB  3.5 TiB   1.7 TiB   10
>> MiB
>> > >>>>> 5.1 GiB  5.6 TiB  38.24  1.10   27      up          osd.14
>> > >>>>> 17    hdd    9.02330   1.00000  9.0 TiB  3.0 TiB   1.2 TiB  6.4
>> MiB
>> > >>>>> 4.1 GiB  6.0 TiB  33.09  0.95   23      up          osd.17
>> > >>>>> 20    hdd    9.02330   1.00000  9.0 TiB  2.8 TiB   1.1 TiB  5.6
>> MiB
>> > >>>>> 3.8 GiB  6.2 TiB  31.55  0.90   20      up          osd.20
>> > >>>>> 23    hdd    9.02330   1.00000  9.0 TiB  2.6 TiB   828 GiB  4.0
>> MiB
>> > >>>>> 3.3 GiB  6.5 TiB  28.32  0.81   23      up          osd.23
>> > >>>>> 26    hdd    9.02330   1.00000  9.0 TiB  2.9 TiB   1.2 TiB  5.8
>> MiB
>> > >>>>> 3.8 GiB  6.1 TiB  32.12  0.92   26      up          osd.26
>> > >>>>> 29    hdd    9.02330   1.00000  9.0 TiB  3.6 TiB   1.8 TiB   10
>> MiB
>> > >>>>> 5.1 GiB  5.4 TiB  39.73  1.14   24      up          osd.29
>> > >>>>> 31    hdd    9.02330   1.00000  9.0 TiB  2.8 TiB   1.1 TiB  5.8
>> MiB
>> > >>>>> 3.7 GiB  6.2 TiB  31.56  0.91   22      up          osd.31
>> > >>>>> 34    hdd    9.02330   1.00000  9.0 TiB  3.3 TiB   1.5 TiB  8.2
>> MiB
>> > >>>>> 4.6 GiB  5.7 TiB  36.29  1.04   23      up          osd.34
>> > >>>>> 37    hdd    9.02330   1.00000  9.0 TiB  3.2 TiB   1.5 TiB  8.2
>> MiB
>> > >>>>> 4.5 GiB  5.8 TiB  35.51  1.02   20      up          osd.37
>> > >>>>> 40    hdd    9.02330   1.00000  9.0 TiB  3.4 TiB   1.7 TiB  9.3
>> MiB
>> > >>>>> 4.9 GiB  5.6 TiB  38.16  1.09   25      up          osd.40
>> > >>>>> 43    hdd    9.02330   1.00000  9.0 TiB  3.4 TiB   1.6 TiB  8.5
>> MiB
>> > >>>>> 4.8 GiB  5.7 TiB  37.19  1.07   29      up          osd.43
>> > >>>>> 46    hdd    9.02330   1.00000  9.0 TiB  3.1 TiB   1.4 TiB  8.4
>> MiB
>> > >>>>> 4.4 GiB  5.9 TiB  34.85  1.00   23      up          osd.46
>> > >>>>>                          TOTAL  433 TiB  151 TiB    67 TiB  364
>> MiB
>> > >>>>> 210 GiB  282 TiB  34.86
>> > >>>>> MIN/MAX VAR: 0.81/1.28  STDDEV: 3.95
>> > >>>>>
>> > >>>>>
>> > >>>>> Michel
>> > >>>>>
>> > >>>>>
>> > >>>>> On Tue, Jan 30, 2024 at 4:18 PM Wesley Dillingham <
>> > >>>>> wes@xxxxxxxxxxxxxxxxx> wrote:
>> > >>>>>
>> > >>>>>> I now concur you should increase the pg_num as a first step for
>> this
>> > >>>>>> cluster. Disable the pg autoscaler for and increase the volumes
>> pool to
>> > >>>>>> pg_num 256. Then likely re-asses and make the next power of 2
>> jump to 512
>> > >>>>>> and probably beyond.
>> > >>>>>>
>> > >>>>>> Keep in mind this is not going to fix your short term deep-scrub
>> > >>>>>> issue in fact it will increase the number of not scrubbed in
>> time PGs until
>> > >>>>>> the pg_num change is complete.  This is because OSDs dont scrub
>> when they
>> > >>>>>> are backfilling.
>> > >>>>>>
>> > >>>>>> I would sit on 256 for a couple weeks and let scrubs happen then
>> > >>>>>> continue past 256.
>> > >>>>>>
>> > >>>>>> with the ultimate target of around 100-200 PGs per OSD which
>> "ceph
>> > >>>>>> osd df tree" will show you in the PGs column.
>> > >>>>>>
>> > >>>>>> Respectfully,
>> > >>>>>>
>> > >>>>>> *Wes Dillingham*
>> > >>>>>> wes@xxxxxxxxxxxxxxxxx
>> > >>>>>> LinkedIn <http://www.linkedin.com/in/wesleydillingham>
>> > >>>>>>
>> > >>>>>>
>> > >>>>>> On Tue, Jan 30, 2024 at 3:16 AM Michel Niyoyita <
>> micou12@xxxxxxxxx>
>> > >>>>>> wrote:
>> > >>>>>>
>> > >>>>>>> Dear team,
>> > >>>>>>>
>> > >>>>>>> below is the output of ceph df command and the ceph version I am
>> > >>>>>>> running
>> > >>>>>>>
>> > >>>>>>>  ceph df
>> > >>>>>>> --- RAW STORAGE ---
>> > >>>>>>> CLASS     SIZE    AVAIL     USED  RAW USED  %RAW USED
>> > >>>>>>> hdd    433 TiB  282 TiB  151 TiB   151 TiB      34.82
>> > >>>>>>> TOTAL  433 TiB  282 TiB  151 TiB   151 TiB      34.82
>> > >>>>>>>
>> > >>>>>>> --- POOLS ---
>> > >>>>>>> POOL                   ID  PGS   STORED  OBJECTS     USED  %USED
>> > >>>>>>> MAX AVAIL
>> > >>>>>>> device_health_metrics   1    1  1.1 MiB        3  3.2 MiB      0
>> > >>>>>>>  73 TiB
>> > >>>>>>> .rgw.root               2   32  3.7 KiB        8   96 KiB      0
>> > >>>>>>>  73 TiB
>> > >>>>>>> default.rgw.log         3   32  3.6 KiB      209  408 KiB      0
>> > >>>>>>>  73 TiB
>> > >>>>>>> default.rgw.control     4   32      0 B        8      0 B      0
>> > >>>>>>>  73 TiB
>> > >>>>>>> default.rgw.meta        5   32    382 B        2   24 KiB      0
>> > >>>>>>>  73 TiB
>> > >>>>>>> volumes                 6  128   21 TiB    5.68M   62 TiB  22.09
>> > >>>>>>>  73 TiB
>> > >>>>>>> images                  7   32  878 GiB  112.50k  2.6 TiB   1.17
>> > >>>>>>>  73 TiB
>> > >>>>>>> backups                 8   32      0 B        0      0 B      0
>> > >>>>>>>  73 TiB
>> > >>>>>>> vms                     9   32  881 GiB  174.30k  2.5 TiB   1.13
>> > >>>>>>>  73 TiB
>> > >>>>>>> testbench              10   32      0 B        0      0 B      0
>> > >>>>>>>  73 TiB
>> > >>>>>>> root@ceph-mon1:~# ceph --version
>> > >>>>>>> ceph version 16.2.11 (3cf40e2dca667f68c6ce3ff5cd94f01e711af894)
>> > >>>>>>> pacific
>> > >>>>>>> (stable)
>> > >>>>>>> root@ceph-mon1:~#
>> > >>>>>>>
>> > >>>>>>> please advise accordingly
>> > >>>>>>>
>> > >>>>>>> Michel
>> > >>>>>>>
>> > >>>>>>> On Mon, Jan 29, 2024 at 9:48 PM Frank Schilder <frans@xxxxxx>
>> wrote:
>> > >>>>>>>
>> > >>>>>>> > You will have to look at the output of "ceph df" and make a
>> > >>>>>>> decision to
>> > >>>>>>> > balance "objects per PG" and "GB per PG". Increase he PG
>> count for
>> > >>>>>>> the
>> > >>>>>>> > pools with the worst of these two numbers most such that it
>> > >>>>>>> balances out as
>> > >>>>>>> > much as possible. If you have pools that see significantly
>> more
>> > >>>>>>> user-IO
>> > >>>>>>> > than others, prioritise these.
>> > >>>>>>> >
>> > >>>>>>> > You will have to find out for your specific cluster, we can
>> only
>> > >>>>>>> give
>> > >>>>>>> > general guidelines. Make changes, run benchmarks, re-evaluate.
>> > >>>>>>> Take the
>> > >>>>>>> > time for it. The better you know your cluster and your users,
>> the
>> > >>>>>>> better
>> > >>>>>>> > the end result will be.
>> > >>>>>>> >
>> > >>>>>>> > Best regards,
>> > >>>>>>> > =================
>> > >>>>>>> > Frank Schilder
>> > >>>>>>> > AIT Risø Campus
>> > >>>>>>> > Bygning 109, rum S14
>> > >>>>>>> >
>> > >>>>>>> > ________________________________________
>> > >>>>>>> > From: Michel Niyoyita <micou12@xxxxxxxxx>
>> > >>>>>>> > Sent: Monday, January 29, 2024 2:04 PM
>> > >>>>>>> > To: Janne Johansson
>> > >>>>>>> > Cc: Frank Schilder; E Taka; ceph-users
>> > >>>>>>> > Subject: Re:  Re: 6 pgs not deep-scrubbed in time
>> > >>>>>>> >
>> > >>>>>>> > This is how it is set , if you suggest to make some changes
>> please
>> > >>>>>>> advises.
>> > >>>>>>> >
>> > >>>>>>> > Thank you.
>> > >>>>>>> >
>> > >>>>>>> >
>> > >>>>>>> > ceph osd pool ls detail
>> > >>>>>>> > pool 1 'device_health_metrics' replicated size 3 min_size 2
>> > >>>>>>> crush_rule 0
>> > >>>>>>> > object_hash rjenkins pg_num 1 pgp_num 1 autoscale_mode on
>> > >>>>>>> last_change 1407
>> > >>>>>>> > flags hashpspool stripe_width 0 pg_num_max 32 pg_num_min 1
>> > >>>>>>> application
>> > >>>>>>> > mgr_devicehealth
>> > >>>>>>> > pool 2 '.rgw.root' replicated size 3 min_size 2 crush_rule 0
>> > >>>>>>> object_hash
>> > >>>>>>> > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change
>> 1393
>> > >>>>>>> flags
>> > >>>>>>> > hashpspool stripe_width 0 application rgw
>> > >>>>>>> > pool 3 'default.rgw.log' replicated size 3 min_size 2
>> crush_rule 0
>> > >>>>>>> > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on
>> > >>>>>>> last_change
>> > >>>>>>> > 1394 flags hashpspool stripe_width 0 application rgw
>> > >>>>>>> > pool 4 'default.rgw.control' replicated size 3 min_size 2
>> > >>>>>>> crush_rule 0
>> > >>>>>>> > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on
>> > >>>>>>> last_change
>> > >>>>>>> > 1395 flags hashpspool stripe_width 0 application rgw
>> > >>>>>>> > pool 5 'default.rgw.meta' replicated size 3 min_size 2
>> crush_rule 0
>> > >>>>>>> > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on
>> > >>>>>>> last_change
>> > >>>>>>> > 1396 flags hashpspool stripe_width 0 pg_autoscale_bias 4
>> > >>>>>>> application rgw
>> > >>>>>>> > pool 6 'volumes' replicated size 3 min_size 2 crush_rule 0
>> > >>>>>>> object_hash
>> > >>>>>>> > rjenkins pg_num 128 pgp_num 128 autoscale_mode on last_change
>> > >>>>>>> 108802 lfor
>> > >>>>>>> > 0/0/14812 flags hashpspool,selfmanaged_snaps stripe_width 0
>> > >>>>>>> application rbd
>> > >>>>>>> >         removed_snaps_queue
>> > >>>>>>> >
>> > >>>>>>>
>> [22d7~3,11561~2,11571~1,11573~1c,11594~6,1159b~f,115b0~1,115b3~1,115c3~1,115f3~1,115f5~e,11613~6,1161f~c,11637~1b,11660~1,11663~2,11673~1,116d1~c,116f5~10,11721~c]
>> > >>>>>>> > pool 7 'images' replicated size 3 min_size 2 crush_rule 0
>> > >>>>>>> object_hash
>> > >>>>>>> > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change
>> 94609
>> > >>>>>>> flags
>> > >>>>>>> > hashpspool,selfmanaged_snaps stripe_width 0 application rbd
>> > >>>>>>> > pool 8 'backups' replicated size 3 min_size 2 crush_rule 0
>> > >>>>>>> object_hash
>> > >>>>>>> > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change
>> 1399
>> > >>>>>>> flags
>> > >>>>>>> > hashpspool stripe_width 0 application rbd
>> > >>>>>>> > pool 9 'vms' replicated size 3 min_size 2 crush_rule 0
>> object_hash
>> > >>>>>>> > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change
>> 108783
>> > >>>>>>> lfor
>> > >>>>>>> > 0/561/559 flags hashpspool,selfmanaged_snaps stripe_width 0
>> > >>>>>>> application rbd
>> > >>>>>>> >         removed_snaps_queue [3fa~1,3fc~3,400~1,402~1]
>> > >>>>>>> > pool 10 'testbench' replicated size 3 min_size 2 crush_rule 0
>> > >>>>>>> object_hash
>> > >>>>>>> > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change
>> 20931
>> > >>>>>>> lfor
>> > >>>>>>> > 0/20931/20929 flags hashpspool stripe_width 0
>> > >>>>>>> >
>> > >>>>>>> >
>> > >>>>>>> > On Mon, Jan 29, 2024 at 2:09 PM Michel Niyoyita <
>> micou12@xxxxxxxxx
>> > >>>>>>> <mailto:
>> > >>>>>>> > micou12@xxxxxxxxx>> wrote:
>> > >>>>>>> > Thank you Janne ,
>> > >>>>>>> >
>> > >>>>>>> > no need of setting some flags like ceph osd set nodeep-scrub
>> ???
>> > >>>>>>> >
>> > >>>>>>> > Thank you
>> > >>>>>>> >
>> > >>>>>>> > On Mon, Jan 29, 2024 at 2:04 PM Janne Johansson <
>> > >>>>>>> icepic.dz@xxxxxxxxx
>> > >>>>>>> > <mailto:icepic.dz@xxxxxxxxx>> wrote:
>> > >>>>>>> > Den mån 29 jan. 2024 kl 12:58 skrev Michel Niyoyita <
>> > >>>>>>> micou12@xxxxxxxxx
>> > >>>>>>> > <mailto:micou12@xxxxxxxxx>>:
>> > >>>>>>> > >
>> > >>>>>>> > > Thank you Frank ,
>> > >>>>>>> > >
>> > >>>>>>> > > All disks are HDDs . Would like to know if I can increase
>> the
>> > >>>>>>> number of
>> > >>>>>>> > PGs
>> > >>>>>>> > > live in production without a negative impact to the
>> cluster. if
>> > >>>>>>> yes which
>> > >>>>>>> > > commands to use .
>> > >>>>>>> >
>> > >>>>>>> > Yes. "ceph osd pool set <poolname> pg_num <number larger than
>> > >>>>>>> before>"
>> > >>>>>>> > where the number usually should be a power of two that leads
>> to a
>> > >>>>>>> > number of PGs per OSD between 100-200.
>> > >>>>>>> >
>> > >>>>>>> > --
>> > >>>>>>> > May the most significant bit of your life be positive.
>> > >>>>>>> >
>> > >>>>>>> _______________________________________________
>> > >>>>>>> ceph-users mailing list -- ceph-users@xxxxxxx
>> > >>>>>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>> > >>>>>>>
>> > >>>>>>
>> > _______________________________________________
>> > ceph-users mailing list -- ceph-users@xxxxxxx
>> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>
>>
>>
>> --
>> May the most significant bit of your life be positive.
>>
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux