Re: I/O freeze while a single node is down.

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



No, no errors about that. I have set noout before it happened, but it
still started recovery. I have added
nobackfill,norebalance,norecover,noscrub,nodeep-scrub once i noticed
it started doing crazy stuff. So recovery I/O stopped but the cluster
can't read any info. Only writes to cache layer.

    cluster cdca2074-4c91-4047-a607-faebcbc1ee17
     health HEALTH_WARN
            2225 pgs degraded
            18 pgs down
            18 pgs peering
            89 pgs stale
            2225 pgs stuck degraded
            18 pgs stuck inactive
            89 pgs stuck stale
            2257 pgs stuck unclean
            2225 pgs stuck undersized
            2225 pgs undersized
            recovery 4180820/11837906 objects degraded (35.317%)
            recovery 24016/11837906 objects misplaced (0.203%)
            12/39 in osds are down
            noout,nobackfill,norebalance,norecover,noscrub,nodeep-scrub
flag(s) set
     monmap e9: 7 mons at {}
            election epoch 170, quorum 0,1,2,3,4,5,6
     osdmap e40290: 40 osds: 27 up, 39 in; 14 remapped pgs
            flags noout,nobackfill,norebalance,norecover,noscrub,nodeep-scrub
      pgmap v39326300: 4096 pgs, 4 pools, 21455 GB data, 5780 kobjects
            42407 GB used, 75772 GB / 115 TB avail
            4180820/11837906 objects degraded (35.317%)
            24016/11837906 objects misplaced (0.203%)
                2136 active+undersized+degraded
                1837 active+clean
                  89 stale+active+undersized+degraded
                  18 down+peering
                  14 active+remapped
                   2 active+clean+scrubbing+deep
  client io 0 B/s rd, 9509 kB/s wr, 3469 op/s

On Tue, Sep 13, 2016 at 1:34 PM, M Ranga Swami Reddy
<swamireddy@xxxxxxxxx> wrote:
> Please check if any osd is nearfull ERR. Can you please share the ceph -s
> o/p?
>
> Thanks
> Swami
>
> On Tue, Sep 13, 2016 at 3:54 PM, Daznis <daznis@xxxxxxxxx> wrote:
>>
>> Hello,
>>
>>
>> I have encountered a strange I/O freeze while rebooting one OSD node
>> for maintenance purpose. It was one of the 3 Nodes in the entire
>> cluster. Before this rebooting or shutting down and entire node just
>> slowed down the ceph, but not completely froze it.
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@xxxxxxxxxxxxxx
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux