A question about rgw.otp pool

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

The autoscale settings of some pools seem to be disabled in some of my clusters.
This problem seems to be caused by overlapping root. My cluster is for RGW
and there are two shadow trees, one is for index (SSD), and the other
is for data (HDD).
This overlapping is caused by the existence of
`ceph-poc-object-store-ssd-index.rgw.otp` pool
which root is not one of the shadow roots, but the default root. In
addition, some other
clusters for RGW, which also have two shadow trees, don't have pools
which names are
postfixed by `rgw.opt`.

Could you tell me how to fix this problem and what is the `...rgw.opt` pool.

Here is the detail.

My cluster uses the following software.

- Ceph: v16.2.7
- Rook: v1.9.6

The result of `ceph osd pool ls`.

```
$ kubectl -n ceph-poc exec deploy/rook-ceph-tools -- ceph osd pool ls
ceph-poc-block-pool
ceph-poc-object-store-ssd-index.rgw.control
ceph-poc-object-store-ssd-index.rgw.meta
ceph-poc-object-store-ssd-index.rgw.log
ceph-poc-object-store-ssd-index.rgw.buckets.index
ceph-poc-object-store-ssd-index.rgw.buckets.non-ec
.rgw.root
ceph-poc-object-store-ssd-index.rgw.buckets.data
ceph-poc-object-store-hdd-index.rgw.control
ceph-poc-object-store-hdd-index.rgw.meta
ceph-poc-object-store-hdd-index.rgw.log
ceph-poc-object-store-hdd-index.rgw.buckets.index
ceph-poc-object-store-hdd-index.rgw.buckets.non-ec
ceph-poc-object-store-hdd-index.rgw.buckets.data
device_health_metrics
ceph-poc-object-store-ssd-index.rgw.otp
```

Some pools are missing in the result of `ceph osd pool autoscale-status`.

```
$ kubectl -n ceph-poc exec deploy/rook-ceph-tools -- ceph osd pool
autoscale-status
POOL                                                  SIZE  TARGET
SIZE  RATE  RAW CAPACITY   RATIO  TARGET RATIO  EFFECTIVE RATIO  BIAS
PG_NUM  NEW PG_NUM  AUTOSCALE  BULK
ceph-poc-object-store-ssd-index.rgw.control             0
   3.0         6144G  0.0000                                  1.0
 8              on         False
ceph-poc-object-store-ssd-index.rgw.meta             3910
   3.0         6144G  0.0000                                  1.0
 8              on         False
ceph-poc-object-store-ssd-index.rgw.log             29328M
   3.0         6144G  0.0140                                  1.0
 8              on         False
ceph-poc-object-store-ssd-index.rgw.buckets.index    4042
   3.0         6144G  0.0000                                  1.0
128           8  off        False
ceph-poc-object-store-ssd-index.rgw.buckets.non-ec      0
   3.0         6144G  0.0000                                  1.0
 8              on         False
.rgw.root                                            9592
   3.0         6144G  0.0000                                  1.0
 8              on         False
device_health_metrics                                8890k
   3.0         6144G  0.0000                                  1.0
32              on         Fals
```

CRUSH rules are as follows.

```
$ kubectl -n ceph-poc exec deploy/rook-ceph-tools -- ceph osd crush
tree --show-shadow
ID    CLASS  WEIGHT     TYPE NAME
  -3    ssd    6.00000  root default~ssd
  -9    ssd    2.00000      zone rack0~ssd
  -8    ssd    1.00000          host 10-69-0-10~ssd
  14    ssd    1.00000              osd.14
 -51    ssd          0          host 10-69-0-22~ssd
...
...
  -2    hdd  781.32037  root default~hdd
  -7    hdd  130.99301      zone rack0~hdd
  -6    hdd          0          host 10-69-0-10~hdd
 -50    hdd   14.55478          host 10-69-0-22~hdd
   8    hdd    7.27739              osd.8
...
  -1         787.32037  root default
  -5         132.99301      zone rack0
  -4           1.00000          host 10-69-0-10
  14    ssd    1.00000              osd.14
 -49          14.55478          host 10-69-0-22
   8    hdd    7.27739              osd.8
...
```

The rule of `...rgw.otp` pools is "replicated_rule".

```
$ kubectl -n ceph-poc exec deploy/rook-ceph-tools -- ceph osd pool get
ceph-poc-object-store-ssd-index.rgw.otp crush_rule
crush_rule: replicated_rule
```

The root of this rule is "default".

```
$ kubectl -n ceph-poc exec deploy/rook-ceph-tools -- ceph osd crush
rule dump replicated_rule
{
    "rule_id": 0,
    "rule_name": "replicated_rule",
    "ruleset": 0,
    "type": 1,
    "min_size": 1,
    "max_size": 10,
    "steps": [
        {
            "op": "take",
            "item": -1,
            "item_name": "default"
        },
        {
            "op": "chooseleaf_firstn",
            "num": 0,
            "type": "host"
        },
        {
            "op": "emit"
        }
    ]
}
```

mgr daemon claims the overlapping root. Here is the log of mgr.

```
...
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.720+0000
7fe64d5d5700  0 [progress INFO root] Processing OSDMap change
175926..175926
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.445+0000
7fe64de96700  0 [pg_autoscaler WARNING root] pool 17 contains an
overlapping root -1... skipping scaling
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.444+0000
7fe64de96700  0 [pg_autoscaler INFO root] Pool 'device_health_metrics'
root_id -3 using 4.139842076256173e-06 of space, bias 1.0, pg target
0.000816928836381218 quantized to 32 (current 32)
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.444+0000
7fe64de96700  0 [pg_autoscaler INFO root] effective_target_ratio 0.0
0.0 0 6597069766656
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.442+0000
7fe64de96700  0 [pg_autoscaler WARNING root] pool 15 contains an
overlapping root -2... skipping scaling
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.441+0000
7fe64de96700  0 [pg_autoscaler WARNING root] pool 14 contains an
overlapping root -2... skipping scaling
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.440+0000
7fe64de96700  0 [pg_autoscaler WARNING root] pool 13 contains an
overlapping root -2... skipping scaling
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.438+0000
7fe64de96700  0 [pg_autoscaler WARNING root] pool 12 contains an
overlapping root -2... skipping scaling
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.437+0000
7fe64de96700  0 [pg_autoscaler WARNING root] pool 11 contains an
overlapping root -2... skipping scaling
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.436+0000
7fe64de96700  0 [pg_autoscaler WARNING root] pool 10 contains an
overlapping root -2... skipping scaling
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.434+0000
7fe64de96700  0 [pg_autoscaler WARNING root] pool 9 contains an
overlapping root -2... skipping scaling
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.433+0000
7fe64de96700  0 [pg_autoscaler INFO root] Pool '.rgw.root' root_id -3
using 4.361936589702964e-09 of space, bias 1.0, pg target
8.607554870347182e-07 quantized to 8 (current 8)
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.433+0000
7fe64de96700  0 [pg_autoscaler INFO root] effective_target_ratio 0.0
0.0 0 6597069766656
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.432+0000
7fe64de96700  0 [pg_autoscaler INFO root] Pool
'ceph-poc-object-store-ssd-index.rgw.buckets.non-ec' root_id -3 using
0.0 of space, bias 1.0, pg target 0.0 quantized to 8 (current 8)
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.432+0000
7fe64de96700  0 [pg_autoscaler INFO root] effective_target_ratio 0.0
0.0 0 6597069766656
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.430+0000
7fe64de96700  0 [pg_autoscaler INFO root] Pool
'ceph-poc-object-store-ssd-index.rgw.buckets.index' root_id -3 using
1.838088792283088e-09 of space, bias 1.0, pg target
3.627161883438627e-07 quantized to 8 (current 128)
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.430+0000
7fe64de96700  0 [pg_autoscaler INFO root] effective_target_ratio 0.0
0.0 0 6597069766656
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.429+0000
7fe64de96700  0 [pg_autoscaler INFO root] Pool
'ceph-poc-object-store-ssd-index.rgw.log' root_id -3 using
0.013985070909257047 of space, bias 1.0, pg target 2.7970141818514094
quantized to 8 (current 8)
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.429+0000
7fe64de96700  0 [pg_autoscaler INFO root] effective_target_ratio 0.0
0.0 0 6597069766656
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.427+0000
7fe64de96700  0 [pg_autoscaler INFO root] Pool
'ceph-poc-object-store-ssd-index.rgw.meta' root_id -3 using
1.7780621419660747e-09 of space, bias 1.0, pg target
3.5561242839321494e-07 quantized to 8 (current 8)
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.427+0000
7fe64de96700  0 [pg_autoscaler INFO root] effective_target_ratio 0.0
0.0 0 6597069766656
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.426+0000
7fe64de96700  0 [pg_autoscaler INFO root] Pool
'ceph-poc-object-store-ssd-index.rgw.control' root_id -3 using 0.0 of
space, bias 1.0, pg target 0.0 quantized to 8 (current 8)
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.426+0000
7fe64de96700  0 [pg_autoscaler INFO root] effective_target_ratio 0.0
0.0 0 6597069766656
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.425+0000
7fe64de96700  0 [pg_autoscaler WARNING root] pool 2 contains an
overlapping root -2... skipping scaling
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.421+0000
7fe64de96700  0 [pg_autoscaler ERROR root] pool 17 has overlapping
roots: {-2, -1}
2022-10-24T10:42:31+09:00 debug 2022-10-24T01:42:31.356+0000
7fe64de96700  0 [pg_autoscaler INFO root] _maybe_adjust
...
```

Regards,
Satoru
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux