health_warn: slow_ops 4 slow ops

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Ceph Community (I'm new here :),

I'm learning Ceph in a Virtual Environment Vagrant/Virtualbox (I understand
this is far from a real environment in several ways, mainly performance,
but I'm ok with that at this point :)

I've 3 nodes, and after few *vagrant halt/up*, when I do *ceph -s*, I got
the following message:

[vagrant@ceph-node1 ~]$ sudo ceph -s
  cluster:
    id:     7f8cb5f0-1989-4ab1-8fb9-d5c08aa96658
    health: *HEALTH_WARN*
            Reduced data availability: 512 pgs inactive
            4 slow ops, oldest one blocked for 1576 sec, daemons
[osd.6,osd.7,osd.8] have slow ops.

  services:
    mon: 3 daemons, quorum ceph-node1,ceph-node2,ceph-node3 (age 7m)
    mgr: ceph-node1(active, since 26m), standbys: ceph-node2, ceph-node3
    osd: 9 osds: 9 up (since 25m), 9 in (since 2d)

  data:
    pools:   1 pools, 512 pgs
    objects: 0 objects, 0 B
    usage:   9.1 GiB used, 162 GiB / 171 GiB avail
    pgs:     100.000% pgs unknown
             512 unknown

Here the output of *ceph health detail*:

[vagrant@ceph-node1 ~]$ sudo ceph health detail
HEALTH_WARN Reduced data availability: 512 pgs inactive; 4 slow ops, oldest
one blocked for 1810 sec, daemons [osd.6,osd.7,osd.8] have slow ops.
PG_AVAILABILITY Reduced data availability: 512 pgs inactive
    pg 2.1cd is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1ce is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1cf is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1d0 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1d1 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1d2 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1d3 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1d4 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1d5 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1d6 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1d7 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1d8 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1d9 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1da is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1db is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1dc is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1dd is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1de is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1df is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1e0 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1e1 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1e2 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1e3 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1e4 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1e5 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1e6 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1e7 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1e8 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1e9 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1ea is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1eb is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1ec is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1ed is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1ee is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1ef is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1f0 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1f1 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1f2 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1f3 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1f4 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1f5 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1f6 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1f7 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1f8 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1f9 is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1fa is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1fb is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1fc is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1fd is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1fe is stuck inactive for 1815.881027, current state unknown, last
acting []
    pg 2.1ff is stuck inactive for 1815.881027, current state unknown, last
acting []
SLOW_OPS 4 slow ops, oldest one blocked for 1810 sec, daemons
[osd.6,osd.7,osd.8] have slow ops.

Do you have any guidance on how to proceed with this? I'm trying to
understand why the cluster is HEALTH_WARN and what I need to do in order to
make it health again.

Thanks!
-- 
Ignacio Ocampo
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux