Re: Ceph Filesystem recovery with intact pools

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



My ceph -s output is this:
  cluster:
    id:     bfe08dcf-aabd-4cac-ac4f-9e56af3df11b
    health: HEALTH_ERR
            1/3 mons down, quorum omicron-m1,omicron-m2
            6 scrub errors
            Possible data damage: 1 pg inconsistent
            Degraded data redundancy: 626702/20558920 objects degraded (3.048%), 32 pgs degraded, 32 pgs undersized
            94 daemons have recently crashed
 
  services:
    mon: 3 daemons, quorum omicron-m1,omicron-m2 (age 8h), out of quorum: omicron-m0
    mgr: omicron-m0(active, since 28h)
    osd: 33 osds: 32 up (since 28h), 32 in (since 28h)
 
  data:
    pools:   8 pools, 736 pgs
    objects: 9.97M objects, 9.9 TiB
    usage:   21 TiB used, 27 TiB / 47 TiB avail
    pgs:     626702/20558920 objects degraded (3.048%)
             702 active+clean
             32  active+undersized+degraded
             1   active+clean+inconsistent
             1   active+clean+scrubbing+deep+repair

and ceph fs status prints nothing.
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux