Re: Massive performance issues

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi folks,

I would also run an iostat -dmx 1 on host 'lenticular' during the fio benchmark just to make sure osd.10 is not being badly hammered with I/Os, which could be capping the cluster's HDD performance due to the very high number of PGs this OSD is involved in.

> 10    hdd   3.63869   1.00000  3.6 TiB  1.9 TiB  1.9 TiB  127 MiB   5.9 GiB  1.7 TiB  52.08  1.19  500      up          osd.10     

Regards,
Frédéric.

----- Le 14 Mar 25, à 8:40, joachim kraftmayer joachim.kraftmayer@xxxxxxxxx a écrit :

> Hi Thomas & Anthony,
> 
> Anthony provided great recommendations.
> 
> ssd read performance:
> I find the total number of pg per ssd osd too low; it can be twice as high.
> 
> hdd read performance
> What makes me a little suspicious is that the maximum throughput of about
> 120 MB/s is exactly the maximum of a 1 Gbit/s connection.
> (I have seen this in the past if the routing is not correct and if you use
> VMs for testing the network could be limited.)
> Can you run your performance test with higher num_jobs, e.g. 16 and post
> the results.
> 
> Regards, Joachim
> 
>  joachim.kraftmayer@xxxxxxxxx
> 
>  www.clyso.com
> 
>  Hohenzollernstr. 27, 80801 Munich
> 
> Utting | HR: Augsburg | HRB: 25866 | USt. ID-Nr.: DE275430677
> 
> 
> 
> Am Fr., 14. März 2025 um 01:53 Uhr schrieb Anthony D'Atri <
> anthony.datri@xxxxxxxxx>:
> 
>>
>> > Hi,
>> >
>> > we are having massive performance issues with our Ceph cluster, and by
>> > now I have no idea how and where to debug further.
>> >
>> > See attachments
>>
>> I see inline text.  If you attached files, those don’t propagate through
>> the list.
>>
>> > for full benchmark logs, relevant excerpts:
>> >
>> > HDD pool:
>> >> read: IOPS=7591, BW=119MiB/s (124MB/s)(34.9GiB/300892msec)
>> >
>> > SSD pool:
>> >> read: IOPS=2007, BW=31.4MiB/s (32.9MB/s)(9422MiB/300334msec)
>> >
>> > Yes, the SSD pool is in fact slower in both IOPS and data rate than the
>> > HDD pool.
>>
>> Making some assumptions based on the below:
>>
>> * Twice as many HDD OSDs as SSD OSDs, so twice the ostensible potential
>> for parallelism
>> * The 5210 ION is QLC from 6 years ago.  Its performance is significantly
>> lower than your Samsungs and especially the 5300s
>> * Your pg_num values are probably fine.
>> * Looks like you are benching against a KRBD mount, so I can’t tell which
>> pool that’s against, or which media your pools are using.  Note that your
>> EC RBD pool is going to be inherently slow with high latency.
>> * An RBD pool will be limited by its slowest OSDs.  Half of your SSDs are
>> 5210s.  So … oh man this is triggering my combinatorics PTSD … many of your
>> PGs in RBD pools using the SSDs will include at least one of them.  The
>> proportion depends on which systems they’re in, which we can’t readily tell
>> from the supplied data.
>> * being respectful of your resources, your systems as described are pretty
>> light CPU-wise for their duties, especially if those collocated VMs are
>> using much at all.
>> * The OSDs on Pileus especially are light on available RAM.  And those
>> include some of the SSDs.
>> * The RAM usage of VMs on the other systems may similarly be starving
>> their OSDs.
>>
>> If you could post `ceph osd metadata` somewhere and send the list a link
>> we can be more specific.
>>
>> Same with `ceph osd crush dump`
>> —aad
>>
>>
>>
>> > Of course this is only one particular benchmark scenario, but
>> > at least it has some numbers and not just “everything feels slow”. I’m
>> > happy to run different benchmarks if required.
>> >
>> > Cluster information:
>> > * 2x10G ethernet to a Cisco Catalyst 3850 on each node
>> > * Debian 11 (bullseye), Linux 5.10
>> > * Ceph 18.2.4 (reef), managed with cephadm
>> > * All but the oldest node serve as VM hypervisors (KVM) as well
>> > * 6 nodes (hostnames appended to compare with OSD information later)
>> >    - 2x AMD EPYC 7232P, 64G RAM (cirrus, nimbus)
>> >    - 1x Intel Xeon E3-1230 v3, 32G RAM (pileus, no VMs)
>> >    - 2x Intel Xeon E5-1620 v4, 64G RAM (lenticular, nacreous)
>> >    - 1x Intel Xeon Silver 4110, 96G RAM (stratus)
>> > * Cluster usage: mostly RBD, some RGW and CephFS
>> > * OSDs:
>> >    - 19 Hitachi Ultrastar 7K4000 2TB
>> >    - 1 WD Ultrastar DC HC310 4TB
>> >    - 3 Hitachi Deskstar 7k2000 2TB
>> >    - 13 Hitachi Ultrastar A7K3000 2TB
>> >    - 8 Micron 5210 ION 2TB
>> >    - 4 Micron 5300 MAX - Mixed Use 2TB
>> >    - 5 Samsung Datacenter PM893 2TB
>> >
>> > You can find various information that I deemed useful below. Please ask
>> > if you would like further information.
>> >
>> > As I said, I don’t really know where to look and what to do, so I would
>> > really appreciate any pointers on how to debug this and improve
>> > performance.
>> >
>> > Of course I have already seen the pg warnings, but I am really not sure
>> > what to adjust in which direction, especially the contradiction between
>> > the MANY_OBJECTS_PER_PG and POOL_TOO_MANY_PGS warnings for the
>> > default.rgw.ec.data pool (there are too many objects per pg, so I
>> > should probably increase pg_num, however it recommends to scale down
>> > from 128 to 32?!).
>> >
>> > Thanks,
>> > Thomas
>> >
>> > # ceph -s
>> >  cluster:
>> >    id:     91688ac0-1b4f-43e3-913d-a844338d9325
>> >    health: HEALTH_WARN
>> >            1 pools have many more objects per pg than average
>> >            3 pools have too many placement groups
>> >
>> >  services:
>> >    mon: 5 daemons, quorum nimbus,lenticular,stratus,cirrus,nacreous (age
>> 5d)
>> >    mgr: lenticular(active, since 3M), standbys: nimbus, stratus,
>> nacreous, cirrus
>> >    mds: 1/1 daemons up, 1 standby
>> >    osd: 53 osds: 53 up (since 2w), 53 in (since 18M)
>> >    rgw: 5 daemons active (5 hosts, 1 zones)
>> >
>> >  data:
>> >    volumes: 1/1 healthy
>> >    pools:   15 pools, 2977 pgs
>> >    objects: 7.20M objects, 15 TiB
>> >    usage:   42 TiB used, 55 TiB / 97 TiB avail
>> >    pgs:     2975 active+clean
>> >             1    active+clean+scrubbing
>> >             1    active+clean+scrubbing+deep
>> >
>> >  io:
>> >    client:   6.5 MiB/s rd, 1.3 MiB/s wr, 158 op/s rd, 63 op/s wr
>> >
>> > # ceph health detail
>> > HEALTH_WARN 1 pools have many more objects per pg than average; 3 pools
>> have too many placement groups
>> > [WRN] MANY_OBJECTS_PER_PG: 1 pools have many more objects per pg than
>> average
>> >    pool default.rgw.ec.data objects per pg (27197) is more than 11.2524
>> times cluster average (2417)
>> > [WRN] POOL_TOO_MANY_PGS: 3 pools have too many placement groups
>> >    Pool default.rgw.ec.data has 128 placement groups, should have 32
>> >    Pool templates has 512 placement groups, should have 256
>> >    Pool rbd.ec has 1024 placement groups, should have 256
>> >
>> > # ceph df
>> > --- RAW STORAGE ---
>> > CLASS    SIZE   AVAIL     USED  RAW USED  %RAW USED
>> > hdd    67 TiB  29 TiB   39 TiB    39 TiB      57.58
>> > ssd    30 TiB  26 TiB  3.7 TiB   3.7 TiB      12.37
>> > TOTAL  97 TiB  55 TiB   42 TiB    42 TiB      43.74
>> >
>> > --- POOLS ---
>> > POOL                        ID   PGS   STORED  OBJECTS     USED  %USED
>> MAX AVAIL
>> > templates                    1   512  4.4 TiB    1.27M   13 TiB  37.46
>>   7.4 TiB
>> > .rgw.root                    9    32  1.4 KiB        4  768 KiB      0
>>   7.4 TiB
>> > default.rgw.control         10    32      0 B        8      0 B      0
>>   7.4 TiB
>> > default.rgw.meta            11    32  7.0 KiB       38  5.1 MiB      0
>>   7.4 TiB
>> > default.rgw.log             12    32  3.5 KiB      208  5.7 MiB      0
>>   7.4 TiB
>> > default.rgw.ec.data         14   128  1.3 TiB    3.48M  2.7 TiB  10.95
>>    13 TiB
>> > default.rgw.buckets.index   15    32   57 MiB       53  171 MiB      0
>>   7.4 TiB
>> > default.rgw.buckets.data    16    32  2.5 GiB   17.85k  8.3 GiB   0.04
>>   7.4 TiB
>> > default.rgw.buckets.non-ec  17    32  1.6 KiB        1  197 KiB      0
>>   7.4 TiB
>> > rbd.ssd                     19   512  872 GiB  259.98k  2.6 TiB   9.46
>>   8.1 TiB
>> > .mgr                        20     1  640 MiB      161  1.9 GiB      0
>>   7.4 TiB
>> > cephfs_metadata             21    32  404 MiB   35.40k  1.2 GiB      0
>>   8.1 TiB
>> > cephfs_data                 22    32   42 GiB   28.09k  128 GiB   0.56
>>   7.4 TiB
>> > rbd                         23   512   12 GiB    6.57k   39 GiB   0.17
>>   7.4 TiB
>> > rbd.ec                      24  1024    8 TiB    2.10M   16 TiB  42.48
>>    11 TiB
>> >
>> > The pool "templates" is named as such for historical reasons. It is
>> > used as RBD VM image storage.
>> >
>> > # ceph osd pool autoscale-status
>> > POOL                          SIZE  TARGET SIZE                RATE  RAW
>> CAPACITY   RATIO  TARGET RATIO  EFFECTIVE RATIO  BIAS  PG_NUM  NEW PG_NUM
>> AUTOSCALE  BULK
>> > .rgw.root                   256.0k                              3.0
>>   68931G  0.0000                                  1.0      32
>> off        False
>> > default.rgw.control             0                               3.0
>>   68931G  0.0000                                  1.0      32
>> off        False
>> > default.rgw.meta             1740k                              3.0
>>   68931G  0.0000                                  1.0      32
>> off        False
>> > default.rgw.log              1932k                              3.0
>>   68931G  0.0000                                  1.0      32
>> off        False
>> > default.rgw.ec.data          1678G               1.6666666269302368
>>   68931G  0.0406                                  1.0     128
>> warn       False
>> > default.rgw.buckets.index   58292k                              3.0
>>   68931G  0.0000                                  1.0      32
>> off        False
>> > default.rgw.buckets.data     2828M                              3.0
>>   68931G  0.0001                                  1.0      32
>> off        False
>> > default.rgw.buckets.non-ec  67159                               3.0
>>   68931G  0.0000                                  1.0      32
>> off        False
>> > .mgr                        639.7M                              3.0
>>   68931G  0.0000                                  1.0       1
>> off        False
>> > cephfs_metadata             405.4M                              3.0
>>   30404G  0.0000                                  4.0      32
>> off        False
>> > cephfs_data                 43733M                              3.0
>>   68931G  0.0019                                  1.0      32
>> off        False
>> > templates                    4541G                              3.0
>>   68931G  0.1977                                  1.0     512
>> warn       True
>> > rbd.ssd                     870.4G                              3.0
>>   30404G  0.0859                                  1.0     512
>> warn       True
>> > rbd                         13336M                              3.0
>>   68931G  0.0006                                  1.0     512
>> off        True
>> > rbd.ec                       8401G                              2.0
>>     68931G  0.2438                                  1.0    1024
>>   warn       True
>> >
>> > # ceph osd pool ls detail
>> > pool 1 'templates' replicated size 3 min_size 2 crush_rule 0 object_hash
>> rjenkins pg_num 512 pgp_num 512 autoscale_mode warn last_change 7495052
>> lfor 0/2976944/7469772 flags hashpspool,selfmanaged_snaps,bulk stripe_width
>> 0 application rbd read_balance_score 1.48
>> > pool 9 '.rgw.root' replicated size 3 min_size 2 crush_rule 0 object_hash
>> rjenkins pg_num 32 pgp_num 32 autoscale_mode off last_change 7467477 flags
>> hashpspool stripe_width 0 application rgw read_balance_score 3.38
>> > pool 10 'default.rgw.control' replicated size 3 min_size 2 crush_rule 0
>> object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode off last_change
>> 7467478 flags hashpspool stripe_width 0 application rgw read_balance_score
>> 3.37
>> > pool 11 'default.rgw.meta' replicated size 3 min_size 2 crush_rule 0
>> object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode off last_change
>> 7467479 flags hashpspool stripe_width 0 application rgw read_balance_score
>> 3.36
>> > pool 12 'default.rgw.log' replicated size 3 min_size 2 crush_rule 0
>> object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode off last_change
>> 7467480 flags hashpspool stripe_width 0 application rgw read_balance_score
>> 4.52
>> > pool 14 'default.rgw.ec.data' erasure profile rgw-video size 5 min_size
>> 4 crush_rule 3 object_hash rjenkins pg_num 128 pgp_num 128 autoscale_mode
>> warn last_change 7494999 lfor 0/5467738/7467533 flags hashpspool
>> stripe_width 12288 application rgw
>> > pool 15 'default.rgw.buckets.index' replicated size 3 min_size 2
>> crush_rule 0 object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode off
>> last_change 7467482 flags hashpspool stripe_width 0 application rgw
>> read_balance_score 4.52
>> > pool 16 'default.rgw.buckets.data' replicated size 3 min_size 2
>> crush_rule 0 object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode off
>> last_change 7467483 flags hashpspool stripe_width 0 application rgw
>> read_balance_score 3.39
>> > pool 17 'default.rgw.buckets.non-ec' replicated size 3 min_size 2
>> crush_rule 0 object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode off
>> last_change 7467484 flags hashpspool stripe_width 0 application rgw
>> read_balance_score 3.38
>> > pool 19 'rbd.ssd' replicated size 3 min_size 2 crush_rule 4 object_hash
>> rjenkins pg_num 512 pgp_num 512 autoscale_mode warn last_change 7495053
>> lfor 0/141997/7467030 flags hashpspool,selfmanaged_snaps,bulk stripe_width
>> 0 application rbd read_balance_score 1.43
>> > pool 20 '.mgr' replicated size 3 min_size 2 crush_rule 0 object_hash
>> rjenkins pg_num 1 pgp_num 1 autoscale_mode off last_change 7480977 flags
>> hashpspool stripe_width 0 pg_num_min 1 application mgr,mgr_devicehealth
>> read_balance_score 37.50
>> > pool 21 'cephfs_metadata' replicated size 3 min_size 2 crush_rule 4
>> object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode off last_change
>> 7467487 lfor 0/0/96365 flags hashpspool stripe_width 0 pg_autoscale_bias 4
>> pg_num_min 16 recovery_priority 5 application cephfs read_balance_score 1.59
>> > pool 22 'cephfs_data' replicated size 3 min_size 2 crush_rule 0
>> object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode off last_change
>> 7467488 lfor 0/0/96365 flags hashpspool,selfmanaged_snaps stripe_width 0
>> application cephfs read_balance_score 4.52
>> > pool 23 'rbd' replicated size 3 min_size 2 crush_rule 0 object_hash
>> rjenkins pg_num 512 pgp_num 512 autoscale_mode off last_change 7467489 lfor
>> 0/0/884114 flags hashpspool,selfmanaged_snaps,bulk stripe_width 0
>> application rbd read_balance_score 1.41
>> > pool 24 'rbd.ec' erasure profile rbd size 4 min_size 3 crush_rule 5
>> object_hash rjenkins pg_num 1024 pgp_num 1024 autoscale_mode warn
>> last_change 7495054 lfor 0/2366959/2366961 flags
>> hashpspool,ec_overwrites,selfmanaged_snaps,bulk stripe_width 8192
>> application rbd
>> >
>> > # ceph osd utilization
>> > avg 192.66
>> > stddev 75.9851 (expected baseline 13.7486)
>> > min osd.48 with 88 pgs (0.456762 * mean)
>> > max osd.10 with 500 pgs (2.59524 * mean)
>> >
>> > # ceph osd df tree
>> > ID   CLASS  WEIGHT    REWEIGHT  SIZE     RAW USE  DATA     OMAP
>>  META      AVAIL    %USE   VAR   PGS  STATUS  TYPE NAME
>> > -13         95.71759         -   97 TiB   42 TiB   42 TiB  4.2 GiB   204
>> GiB   55 TiB  43.74  1.00    -          root default
>> > -16         14.15759         -   15 TiB  8.2 TiB  8.2 TiB  713 MiB    29
>> GiB  6.4 TiB  56.36  1.29    -              host cirrus
>> >  2    hdd   1.76970   1.00000  1.8 TiB  1.1 TiB  1.0 TiB   85 MiB   4.9
>> GiB  784 GiB  57.94  1.32  223      up          osd.2
>> >  5    hdd   1.76970   1.00000  1.8 TiB  1.0 TiB  1.0 TiB   50 MiB   3.2
>> GiB  821 GiB  55.95  1.28  223      up          osd.5
>> >  8    hdd   1.76970   1.00000  1.8 TiB  1.0 TiB  1.0 TiB   88 MiB   2.9
>> GiB  804 GiB  56.86  1.30  229      up          osd.8
>> > 11    hdd   1.76970   1.00000  1.8 TiB  963 GiB  959 GiB  236 MiB   3.6
>> GiB  900 GiB  51.66  1.18  240      up          osd.11
>> > 13    hdd   1.76970   1.00000  1.8 TiB  1.0 TiB  1.0 TiB   61 MiB   4.7
>> GiB  823 GiB  55.81  1.28  220      up          osd.13
>> > 15    hdd   1.76970   1.00000  1.8 TiB  1.1 TiB  1.1 TiB   52 MiB   3.1
>> GiB  762 GiB  59.13  1.35  229      up          osd.15
>> > 18    hdd   1.76970   1.00000  1.8 TiB  1.1 TiB  1.0 TiB   85 MiB   2.7
>> GiB  787 GiB  57.76  1.32  228      up          osd.18
>> > 19    hdd   1.76970   1.00000  1.8 TiB  1.0 TiB  1.0 TiB   54 MiB   3.4
>> GiB  824 GiB  55.79  1.28  219      up          osd.19
>> > -18         15.93417         -   16 TiB  6.0 TiB  6.0 TiB  663 MiB    23
>> GiB   10 TiB  37.15  0.85    -              host lenticular
>> >  0    hdd   1.76970   1.00000  1.8 TiB  1.0 TiB  1.0 TiB   88 MiB   3.2
>> GiB  796 GiB  57.26  1.31  229      up          osd.0
>> >  4    hdd   1.76970   1.00000  1.8 TiB  1.1 TiB  1.1 TiB   42 MiB   6.0
>> GiB  753 GiB  59.57  1.36  231      up          osd.4
>> >  6    hdd   1.76970   1.00000  1.8 TiB  1.1 TiB  1.1 TiB   54 MiB   3.5
>> GiB  758 GiB  59.31  1.36  237      up          osd.6
>> > 10    hdd   3.63869   1.00000  3.6 TiB  1.9 TiB  1.9 TiB  127 MiB   5.9
>> GiB  1.7 TiB  52.08  1.19  500      up          osd.10
>> > 36    ssd   1.74660   1.00000  1.7 TiB  229 GiB  228 GiB   92 MiB   959
>> MiB  1.5 TiB  12.79  0.29  103      up          osd.36
>> > 37    ssd   1.74660   1.00000  1.7 TiB  223 GiB  222 GiB   97 MiB  1006
>> MiB  1.5 TiB  12.49  0.29   98      up          osd.37
>> > 38    ssd   1.74660   1.00000  1.7 TiB  223 GiB  222 GiB   61 MiB   1.0
>> GiB  1.5 TiB  12.46  0.28   98      up          osd.38
>> > 39    ssd   1.74660   1.00000  1.7 TiB  221 GiB  219 GiB  102 MiB   1.4
>> GiB  1.5 TiB  12.35  0.28   98      up          osd.39
>> > -17         22.89058         -   23 TiB  9.5 TiB  9.4 TiB  1.1 GiB    45
>> GiB   14 TiB  40.59  0.93    -              host nacreous
>> >  1    hdd   1.76970   1.00000  1.8 TiB  1.0 TiB  1.0 TiB  124 MiB   6.0
>> GiB  818 GiB  56.07  1.28  229      up          osd.1
>> >  3    hdd   1.76970   1.00000  1.8 TiB  1.1 TiB  1.1 TiB   54 MiB   3.1
>> GiB  780 GiB  58.11  1.33  222      up          osd.3
>> >  7    hdd   1.76970   1.00000  1.8 TiB  1.1 TiB  1.1 TiB   48 MiB   3.1
>> GiB  773 GiB  58.48  1.34  237      up          osd.7
>> >  9    hdd   1.76970   1.00000  1.8 TiB  1.1 TiB  1.1 TiB   50 MiB   4.9
>> GiB  766 GiB  58.88  1.35  223      up          osd.9
>> > 14    hdd   1.76970   1.00000  1.8 TiB  1.0 TiB  1.0 TiB   93 MiB   3.3
>> GiB  812 GiB  56.43  1.29  224      up          osd.14
>> > 16    hdd   1.76970   1.00000  1.8 TiB  1.0 TiB  1.0 TiB   43 MiB   5.0
>> GiB  792 GiB  57.51  1.31  235      up          osd.16
>> > 17    hdd   1.76970   1.00000  1.8 TiB  1.1 TiB  1.1 TiB   68 MiB   4.2
>> GiB  777 GiB  58.31  1.33  227      up          osd.17
>> > 20    hdd   1.76970   1.00000  1.8 TiB  1.1 TiB  1.1 TiB   66 MiB   5.9
>> GiB  780 GiB  58.15  1.33  231      up          osd.20
>> > 48    ssd   1.74660   1.00000  1.7 TiB  213 GiB  212 GiB   94 MiB   1.3
>> GiB  1.5 TiB  11.92  0.27   88      up          osd.48
>> > 49    ssd   1.74660   1.00000  1.7 TiB  215 GiB  212 GiB   94 MiB   2.5
>> GiB  1.5 TiB  12.01  0.27   90      up          osd.49
>> > 50    ssd   1.74660   1.00000  1.7 TiB  214 GiB  212 GiB   95 MiB   2.1
>> GiB  1.5 TiB  11.99  0.27   92      up          osd.50
>> > 51    ssd   1.74660   1.00000  1.7 TiB  217 GiB  215 GiB  144 MiB   1.9
>> GiB  1.5 TiB  12.13  0.28   94      up          osd.51
>> > 52    ssd   1.74660   1.00000  1.7 TiB  214 GiB  212 GiB  144 MiB   2.0
>> GiB  1.5 TiB  11.95  0.27   94      up          osd.52
>> > -19          6.98639         -  7.0 TiB  895 GiB  889 GiB  345 MiB   5.6
>> GiB  6.1 TiB  12.51  0.29    -              host nimbus
>> > 44    ssd   1.74660   1.00000  1.7 TiB  227 GiB  224 GiB  148 MiB   2.3
>> GiB  1.5 TiB  12.67  0.29   98      up          osd.44
>> > 45    ssd   1.74660   1.00000  1.7 TiB  222 GiB  221 GiB   68 MiB  1021
>> MiB  1.5 TiB  12.39  0.28   93      up          osd.45
>> > 46    ssd   1.74660   1.00000  1.7 TiB  223 GiB  222 GiB   64 MiB   972
>> MiB  1.5 TiB  12.48  0.29   97      up          osd.46
>> > 47    ssd   1.74660   1.00000  1.7 TiB  224 GiB  222 GiB   64 MiB   1.3
>> GiB  1.5 TiB  12.52  0.29   95      up          osd.47
>> > -15         21.19368         -   22 TiB  9.3 TiB  9.3 TiB  848 MiB    56
>> GiB   12 TiB  43.35  0.99    -              host pileus
>> > 12    hdd   1.76970   1.00000  1.8 TiB  1.0 TiB  1.0 TiB   54 MiB   8.7
>> GiB  790 GiB  57.61  1.32  231      up          osd.12
>> > 21    hdd   1.76970   1.00000  1.8 TiB  1.1 TiB  1.1 TiB   46 MiB   6.1
>> GiB  765 GiB  58.94  1.35  236      up          osd.21
>> > 22    hdd   1.76970   1.00000  1.8 TiB  1.0 TiB  1.0 TiB   60 MiB   6.2
>> GiB  810 GiB  56.52  1.29  228      up          osd.22
>> > 23    hdd   1.76970   1.00000  1.8 TiB  1.1 TiB  1.1 TiB   83 MiB   8.0
>> GiB  779 GiB  58.16  1.33  230      up          osd.23
>> > 24    hdd   1.76970   1.00000  1.8 TiB  1.1 TiB  1.1 TiB   44 MiB   5.5
>> GiB  777 GiB  58.29  1.33  229      up          osd.24
>> > 25    hdd   1.76970   1.00000  1.8 TiB  1.0 TiB  1.0 TiB   73 MiB   5.6
>> GiB  799 GiB  57.09  1.31  230      up          osd.25
>> > 26    hdd   1.76970   1.00000  1.8 TiB  1.0 TiB  1.0 TiB   63 MiB   5.4
>> GiB  797 GiB  57.22  1.31  239      up          osd.26
>> > 27    hdd   1.81940   1.00000  1.8 TiB  1.1 TiB  1.1 TiB   99 MiB   7.1
>> GiB  720 GiB  61.36  1.40  247      up          osd.27
>> > 40    ssd   1.74660   1.00000  1.7 TiB  220 GiB  219 GiB   66 MiB  1013
>> MiB  1.5 TiB  12.31  0.28   96      up          osd.40
>> > 41    ssd   1.74660   1.00000  1.7 TiB  221 GiB  220 GiB   63 MiB   981
>> MiB  1.5 TiB  12.36  0.28   97      up          osd.41
>> > 42    ssd   1.74660   1.00000  1.7 TiB  228 GiB  227 GiB   98 MiB   999
>> MiB  1.5 TiB  12.73  0.29  101      up          osd.42
>> > 43    ssd   1.74660   1.00000  1.7 TiB  227 GiB  226 GiB   99 MiB   866
>> MiB  1.5 TiB  12.67  0.29  100      up          osd.43
>> > -14         14.55518         -   15 TiB  8.6 TiB  8.5 TiB  582 MiB    45
>> GiB  6.0 TiB  59.02  1.35    -              host stratus
>> > 28    hdd   1.81940   1.00000  1.8 TiB  1.1 TiB  1.1 TiB   54 MiB   5.5
>> GiB  756 GiB  59.41  1.36  239      up          osd.28
>> > 29    hdd   1.81940   1.00000  1.8 TiB  1.1 TiB  1.1 TiB   44 MiB   4.7
>> GiB  759 GiB  59.27  1.35  230      up          osd.29
>> > 30    hdd   1.81940   1.00000  1.8 TiB  1.1 TiB  1.1 TiB   68 MiB   5.0
>> GiB  753 GiB  59.57  1.36  232      up          osd.30
>> > 31    hdd   1.81940   1.00000  1.8 TiB  1.1 TiB  1.1 TiB   64 MiB   6.2
>> GiB  778 GiB  58.26  1.33  229      up          osd.31
>> > 32    hdd   1.81940   1.00000  1.8 TiB  1.0 TiB  1.0 TiB   85 MiB   5.0
>> GiB  797 GiB  57.24  1.31  226      up          osd.32
>> > 33    hdd   1.81940   1.00000  1.8 TiB  1.1 TiB  1.1 TiB  112 MiB   4.1
>> GiB  749 GiB  59.81  1.37  235      up          osd.33
>> > 34    hdd   1.81940   1.00000  1.8 TiB  1.1 TiB  1.1 TiB   46 MiB   8.5
>> GiB  770 GiB  58.66  1.34  238      up          osd.34
>> > 35    hdd   1.81940   1.00000  1.8 TiB  1.1 TiB  1.1 TiB  109 MiB   6.2
>> GiB  747 GiB  59.92  1.37  244      up          osd.35
>> >                         TOTAL   97 TiB   42 TiB   42 TiB  4.2 GiB   204
>> GiB   55 TiB  43.74
>> > MIN/MAX VAR: 0.27/1.40  STDDEV: 21.24
>> >
>> > # ceph balancer status
>> > {
>> >    "active": true,
>> >    "last_optimize_duration": "0:00:00.040070",
>> >    "last_optimize_started": "Thu Mar 13 14:59:05 2025",
>> >    "mode": "upmap",
>> >    "no_optimization_needed": true,
>> >    "optimize_result": "Unable to find further optimization, or pool(s)
>> pg_num is decreasing, or distribution is already perfect",
>> >    "plans": []
>> > }
>> >
>> > --
>> > Fachschaft I/1 Mathematik/Physik/Informatik der RWTH Aachen Thomas
>> > Schneider Campus Mitte: Augustinerbach 2a, 52062 Aachen Telefon: +49
>> > 241 80 94506 Informatikzentrum: Ahornstraße 55, Raum 4U17, 52074 Aachen
>> > Telefon: +49 241 80 26741 https://www.fsmpi.rwth-aachen.de
>> > # fio --ioengine=libaio --direct=1 --bs=16384 --iodepth=128
>> --rw=randread --norandommap --size=20G --numjobs=1 --runtime=300
>> --time_based --name=/dev/rbd/rbd/test
>> > /dev/rbd/rbd/test: (g=0): rw=randread, bs=(R) 16.0KiB-16.0KiB, (W)
>> 16.0KiB-16.0KiB, (T) 16.0KiB-16.0KiB, ioengine=libaio, iodepth=128
>> > fio-3.25
>> > Starting 1 process
>> > Jobs: 1 (f=1): [r(1)][100.0%][r=35.0MiB/s][r=2241 IOPS][eta 00m:00s]
>> > /dev/rbd/rbd/test: (groupid=0, jobs=1): err= 0: pid=3832471: Thu Jan 30
>> 19:55:46 2025
>> >  read: IOPS=7591, BW=119MiB/s (124MB/s)(34.9GiB/300892msec)
>> >    slat (usec): min=2, max=665, avg= 8.93, stdev= 5.57
>> >    clat (usec): min=65, max=5330.5k, avg=16850.13, stdev=184056.88
>> >     lat (usec): min=82, max=5330.5k, avg=16859.23, stdev=184056.97
>> >    clat percentiles (usec):
>> >     |  1.00th=[    147],  5.00th=[    180], 10.00th=[    202],
>> >     | 20.00th=[    237], 30.00th=[    269], 40.00th=[    297],
>> >     | 50.00th=[    330], 60.00th=[    367], 70.00th=[    420],
>> >     | 80.00th=[    523], 90.00th=[   1876], 95.00th=[   6128],
>> >     | 99.00th=[ 362808], 99.50th=[1098908], 99.90th=[3506439],
>> >     | 99.95th=[4143973], 99.99th=[4731175]
>> >   bw (  KiB/s): min=  768, max=478432, per=100.00%, avg=121819.01,
>> stdev=95962.09, samples=600
>> >   iops        : min=   48, max=29902, avg=7613.70, stdev=5997.62,
>> samples=600
>> >  lat (usec)   : 100=0.01%, 250=24.52%, 500=53.85%, 750=8.61%, 1000=1.56%
>> >  lat (msec)   : 2=1.84%, 4=3.46%, 10=2.34%, 20=1.12%, 50=0.91%
>> >  lat (msec)   : 100=0.40%, 250=0.31%, 500=0.13%, 750=0.18%, 1000=0.21%
>> >  lat (msec)   : 2000=0.34%, >=2000=0.20%
>> >  cpu          : usr=2.33%, sys=10.06%, ctx=1687235, majf=0, minf=523
>> >  IO depths    : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%,
>> >=64=100.0%
>> >     submit    : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%,
>> >=64=0.0%
>> >     complete  : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%,
>> >=64=0.1%
>> >     issued rwts: total=2284212,0,0,0 short=0,0,0,0 dropped=0,0,0,0
>> >     latency   : target=0, window=0, percentile=100.00%, depth=128
>> >
>> > Run status group 0 (all jobs):
>> >   READ: bw=119MiB/s (124MB/s), 119MiB/s-119MiB/s (124MB/s-124MB/s),
>> io=34.9GiB (37.4GB), run=300892-300892msec
>> > <fio-ssd.txt>
>> > _______________________________________________
>> > ceph-users mailing list -- ceph-users@xxxxxxx
>> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@xxxxxxx
>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux