Re: Stuck in remapped state?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Update on "ceph-s". A machine was in the process of crashing when I
took the original snapshot. Here it is after the reboot:

[root@dell02 ~]# ceph -s
  cluster:
    id:     278fcd86-0861-11ee-a7df-9c5c8e86cf8f
    health: HEALTH_WARN
            1 filesystem is degraded
            25 client(s) laggy due to laggy OSDs
 
  services:
    mon: 3 daemons, quorum dell02,www7,ceph03 (age 8m)
    mgr: ceph08.tlocfi(active, since 81m), standbys: www7.rxagfn,
dell02.odtbqw
    mds: 1/1 daemons up, 2 standby
    osd: 7 osds: 7 up (since 12h), 7 in (since 19h); 308 remapped pgs
    rgw: 2 daemons active (2 hosts, 1 zones)
 
  data:
    volumes: 0/1 healthy, 1 recovering
    pools:   22 pools, 681 pgs
    objects: 125.10k objects, 36 GiB
    usage:   91 GiB used, 759 GiB / 850 GiB avail
    pgs:     47772/369076 objects misplaced (12.944%)
             373 active+clean
             308 active+clean+remapped
 
  io:
    client:   170 B/s rd, 0 op/s rd, 0 op/s wr

On Sat, 2024-07-27 at 11:31 -0400, Tim Holloway wrote:
> I was in the  middle of tuning my OSDs when lightning blew me off the
> Internet. Had to wait 5 days for my ISP to send a tech and replace a
> fried cable. In the mean time, among other things. I had some serious
> time drift between servers thanks to the OS upgrades replacing NTP
> with
> chrony and me not having thought to re-establish a master in-house
> timeserver.
> 
> Ceph tried really hard to keep up with all that, but eventually it
> was
> just too much. Now I've got an offline filesystem and apparently it's
> stuck trying to get back online again.
> 
> The forensics:
> [ceph: root@www7 /]# ceph -s
>   cluster:
>     id:     278fcd86-0861-11ee-a7df-9c5c8e86cf8f
>     health: HEALTH_WARN
>             failed to probe daemons or devices
>             1 filesystem is degraded
>             1/3 mons down, quorum www7,ceph03
>  
>   services:
>     mon: 3 daemons, quorum www7,ceph03 (age 2m), out of quorum:
> dell02
>     mgr: ceph08.tlocfi(active, since 58m), standbys: dell02.odtbqw,
> www7.rxagfn
>     mds: 1/1 daemons up, 1 standby
>     osd: 7 osds: 7 up (since 12h), 7 in (since 18h); 308 remapped pgs
>     rgw: 2 daemons active (2 hosts, 1 zones)
>  
>   data:
>     volumes: 0/1 healthy, 1 recovering
>     pools:   22 pools, 681 pgs
>     objects: 125.10k objects, 36 GiB
>     usage:   91 GiB used, 759 GiB / 850 GiB avail
>     pgs:     47772/369076 objects misplaced (12.944%)
>              373 active+clean
>              308 active+clean+remapped
>  
>   io:
>     client:   170 B/s rd, 0 op/s rd, 0 op/s wr
> 
> [ceph: root@www7 /]# ceph health detail
> HEALTH_WARN 1 filesystem is degraded; 25 client(s) laggy due to laggy
> OSDs
> [WRN] FS_DEGRADED: 1 filesystem is degraded
>     fs ceefs is degraded
> [WRN] MDS_CLIENTS_LAGGY: 25 client(s) laggy due to laggy OSDs
>     mds.ceefs.www7.drnuyi(mds.0): Client 14019719 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14124385 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14144243 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14144375 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14224103 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14224523 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14234194 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14234545 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14236841 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14237837 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14238536 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14244124 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14264236 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14266870 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14294170 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14294434 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14296012 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14304212 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14316057 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14318379 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14325518 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14328956 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14334283 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14336104 is laggy; not
> evicted
> because some OSD(s) is/are laggy
>     mds.ceefs.www7.drnuyi(mds.0): Client 14374237 is laggy; not
> evicted
> because some OSD(s) is/are laggy
> 
> [ceph: root@www7 /]# ceph osd tree
> ID   CLASS  WEIGHT   TYPE NAME        STATUS  REWEIGHT  PRI-AFF
>  -1         2.79994  root default                              
> -25         0.15999      host ceph01                           
>   1    hdd  0.15999          osd.1        up   0.15999  1.00000
> -28         1.15999      host ceph03                           
>   3    hdd  0.15999          osd.3        up   0.15999  1.00000
>   5    hdd  1.00000          osd.5        up   1.00000  1.00000
>  -9         0.15999      host ceph06                           
>   2    hdd  0.15999          osd.2        up   0.15999  1.00000
>  -3         0.15999      host ceph07                           
>   6    hdd  0.15999          osd.6        up   0.15999  1.00000
>  -6         1.00000      host ceph08                           
>   4    hdd  1.00000          osd.4        up   1.00000  1.00000
>  -7         0.15999      host www7                             
>   0    hdd  0.15999          osd.0        up   0.15999  1.00000
> 
> [ceph: root@www7 /]# ceph pg stat
> 681 pgs: 373 active+clean, 308 active+clean+remapped; 36 GiB data, 91
> GiB used, 759 GiB / 850 GiB avail; 255 B/s rd, 0 op/s; 47772/369073
> objects misplaced (12.944%)
> 
> 
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux