Re: Stuck MDSs behind in trimming

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



After some more digging, all three MDS enter state up:rejoin but don't move
on from there when restarting.

Also, MDS 0 (not the one with a trimming problem) consistently has
mds.0.cache  failed to open ino 0x101 err -116/0
mds.0.cache  failed to open ino 0x102 err -116/0

in the log when restarting.

On Thu, Jul 8, 2021 at 6:29 AM Zachary Ulissi <zulissi@xxxxxxxxx> wrote:

> We're running a rook-ceph cluster that has gotten stuck in "1 MDSs behind
> on trimming".
>
> * 1 filesystem, three active MDS servers each with standby
> * Quite a few files (20M objects), daily snapshots. This might be a
> problem?
> * Ceph pacific 16.2.4
>
> * `ceph health detail` doesn't provide much help (see below)
> * num_segments is very slowly increasing over time
> * Restarting all of the MDSs returns to the same point.
> * moderate CPU usage for each MDS server (~30% for the stuck one, ~80% of
> a core for the others)
> * logs for the stuck MDS looks clean, it hits rejoin_joint_start then
> standard 'updating MDS map to version XXX" messages
> * `ceph daemon mds.x ops` shows no active ops on each of the MDS servers
> * `mds_log_max_segments` is set to 128, setting to a higher number causes
> the warning to go away, but the filesystem remains degraded, and setting it
> back to 128 shows num_segments has not changed.
> * I've tried playing around with other MDS settings based on various posts
> on this list and elsewhere, to no avail
> * `cephfs-journal-tool journal inspect` for each rank says journal
> integrity is fine.
>
> Something similar happened last week and (probably by accident by
> removing/adding nodes?) I got the MDSs to start recovering and the
> filesystem went back to healthy.
>
> I'm at a bit of a loss for what else to try.
>
> Thanks!
> Zack
>
>
> `ceph health detail`
> HEALTH_WARN mons are allowing insecure global_id reclaim; 1 filesystem is
> degraded; 1 MDSs behind on trimming; mon x is low on available space
> [WRN] AUTH_INSECURE_GLOBAL_ID_RECLAIM_ALLOWED: mons are allowing insecure
> global_id reclaim
>     mon.x has auth_allow_insecure_global_id_reclaim set to true
>     mon.ad has auth_allow_insecure_global_id_reclaim set to true
>     mon.af has auth_allow_insecure_global_id_reclaim set to true
> [WRN] FS_DEGRADED: 1 filesystem is degraded
>     fs myfs is degraded
> [WRN] MDS_TRIM: 1 MDSs behind on trimming
>     mds.myfs-d(mds.2): Behind on trimming (340/128) max_segments: 128,
> num_segments: 340
> [WRN] MON_DISK_LOW: mon x is low on available space
>     mon.x has 22% avail
>
> `ceph config get mds`
> WHO     MASK  LEVEL     OPTION                              VALUE        RO
> global        basic     log_file                                         *
> global        basic     log_to_file                         false
> mds           basic     mds_cache_memory_limit              17179869184
> mds           advanced  mds_cache_trim_decay_rate           1.000000
> mds           advanced  mds_cache_trim_threshold            1048576
> mds           advanced  mds_log_max_segments                128
> mds           advanced  mds_recall_max_caps                 5000
> mds           advanced  mds_recall_max_decay_rate           2.500000
> global        advanced  mon_allow_pool_delete               true
> global        advanced  mon_allow_pool_size_one             true
> global        advanced  mon_cluster_log_file
> global        advanced  mon_pg_warn_min_per_osd             0
> global        advanced  osd_pool_default_pg_autoscale_mode  on
> global        advanced  osd_scrub_auto_repair               true
> global        advanced  rbd_default_features                3
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux