Issue with Nautilus upgrade from Luminous

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Dear All,

We have 13  Ceph clusters and we started upgrading one by one from Luminous
to Nautilus. Post upgrade started fixing the warning alerts and had issues
setting "*ceph config set mon mon_crush_min_required_version firefly" *yielded
no results.  Updated the mon config and restart the daemons the warning
didn't go away

I have also tried to set it to hammer and no use.  The warning is still
there.  Do you have any recommendations?  I thought of changing it to
hammer so I can use straw2 but I was stuck with warning message.  I have
also bounced the nodes and the issue remains the same.

Please review and share your inputs.

  cluster:
    id:     xxxxxxxxxxx
    health: HEALTH_WARN
            crush map has legacy tunables (require firefly, min is hammer)
            1 pools have many more objects per pg than average
            15252 pgs not deep-scrubbed in time
            21399 pgs not scrubbed in time
            clients are using insecure global_id reclaim
            mons are allowing insecure global_id reclaim
            3 monitors have not enabled msgr2


ceph daemon mon.$(hostname -s) config show |grep -i
mon_crush_min_required_version
    "mon_crush_min_required_version": "firefly",

ceph osd crush show-tunables
{
    "choose_local_tries": 0,
    "choose_local_fallback_tries": 0,
    "choose_total_tries": 50,
    "chooseleaf_descend_once": 1,
    "chooseleaf_vary_r": 1,
    "chooseleaf_stable": 0,
    "straw_calc_version": 1,
    "allowed_bucket_algs": 22,
    "profile": "firefly",
    "optimal_tunables": 0,
    "legacy_tunables": 0,
    "minimum_required_version": "firefly",
    "require_feature_tunables": 1,
    "require_feature_tunables2": 1,
    "has_v2_rules": 0,
    "require_feature_tunables3": 1,
    "has_v3_rules": 0,
    "has_v4_buckets": 0,
    "require_feature_tunables5": 0,
    "has_v5_rules": 0
}

ceph config dump
WHO   MASK LEVEL    OPTION                         VALUE   RO
  mon      advanced mon_crush_min_required_version firefly *

ceph versions
{
    "mon": {
        "ceph version 14.2.22 (ca74598065096e6fcbd8433c8779a2be0c889351)
nautilus (stable)": 3
    },
    "mgr": {
        "ceph version 14.2.22 (ca74598065096e6fcbd8433c8779a2be0c889351)
nautilus (stable)": 3
    },
    "osd": {
        "ceph version 14.2.21 (5ef401921d7a88aea18ec7558f7f9374ebd8f5a6)
nautilus (stable)": 549,
        "ceph version 14.2.22 (ca74598065096e6fcbd8433c8779a2be0c889351)
nautilus (stable)": 226
    },
    "mds": {},
    "rgw": {
        "ceph version 14.2.22 (ca74598065096e6fcbd8433c8779a2be0c889351)
nautilus (stable)": 2
    },
    "overall": {
        "ceph version 14.2.21 (5ef401921d7a88aea18ec7558f7f9374ebd8f5a6)
nautilus (stable)": 549,
        "ceph version 14.2.22 (ca74598065096e6fcbd8433c8779a2be0c889351)
nautilus (stable)": 234
    }
}

ceph -s
  cluster:
    id:    xxxxxxxxxxxxxxxxxx
    health: HEALTH_WARN
            crush map has legacy tunables (require firefly, min is hammer)
            1 pools have many more objects per pg than average
            13811 pgs not deep-scrubbed in time
            19994 pgs not scrubbed in time
            clients are using insecure global_id reclaim
            mons are allowing insecure global_id reclaim
            3 monitors have not enabled msgr2

  services:
    mon: 3 daemons, quorum
pistoremon-ho-c01,pistoremon-ho-c02,pistoremon-ho-c03 (age 24s)
    mgr: pistoremon-ho-c02(active, since 2m), standbys: pistoremon-ho-c01,
pistoremon-ho-c03
    osd: 800 osds: 775 up (since 105m), 775 in
    rgw: 2 daemons active (pistorergw-ho-c01, pistorergw-ho-c02)

  task status:

  data:
    pools:   28 pools, 27336 pgs
    objects: 107.19M objects, 428 TiB
    usage:   1.3 PiB used, 1.5 PiB / 2.8 PiB avail
    pgs:     27177 active+clean
             142   active+clean+scrubbing+deep
             17    active+clean+scrubbing

  io:
    client:   220 MiB/s rd, 1.9 GiB/s wr, 7.07k op/s rd, 25.42k op/s wr

-- 
Regards,
Suresh
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux