Re: pgs stuck inactive and unclean, too feww PGs per OSD

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,
        I've removed the rbd pool and created it again. It picked up my default settings but there are still some problems.
        After running "sudo ceph -s", the output is as follow:
 
    cluster 0b9b05db-98fe-49e6-b12b-1cce0645c015
     health HEALTH_WARN
            512 pgs stuck inactive
            512 pgs stuck unclean
     monmap e1: 1 mons at {monitorOne=192.168.1.153:6789/0}
            election epoch 1, quorum 0 monitorOne
     osdmap e62: 8 osds: 8 up, 8 in
      pgmap v219: 512 pgs, 1 pools, 0 bytes data, 0 objects
            8460 MB used, 4162 GB / 4171 GB avail
                 512 creating

Ceph stucks in creating the pgs forever. Those pgs are stuck in inactive and unclean. And the Ceph pg query hangs forever.
I googled this problem and didn't get a clue.
Is there anything I missed?
Any idea to help me?

--
Zhen Wang


At 2015-10-07 13:05:51, "Christian Balzer" <chibi@xxxxxxx> wrote: > >Hello, >On Wed, 7 Oct 2015 12:57:58 +0800 (CST) wikison wrote: > >This is a very old bug, misfeature. >And creeps up every week or so here, google is your friend. > >> Hi, >> I have a cluster of one monitor and eight OSDs. These OSDs are running >> on four hosts(each host has two OSDs). When I set up everything and >> started Ceph, I got this: esta@monitorOne:~$ sudo ceph -s [sudo] >> password for esta: cluster 0b9b05db-98fe-49e6-b12b-1cce0645c015 >> health HEALTH_WARN >> 64 pgs stuck inactive >> 64 pgs stuck unclean >> too few PGs per OSD (8 < min 30) > >Those 3 lines tell you pretty much all there is wrong. >You did (correctly) set the defaul pg and pgp nums to something sensible >(512) in your ceph.conf. >Unfortunately when creating the initial pool (rbd) it still ignores those >settings. > >You could try to increase those for your pool, which may or may not work. > >The easier and faster way is to remove the rbd pool and create it again. >This should pick up your default settings. > >Christian > >> monmap e1: 1 mons at {monitorOne=192.168.1.153:6789/0} >> election epoch 1, quorum 0 monitorOne >> osdmap e58: 8 osds: 8 up, 8 in >> pgmap v191: 64 pgs, 1 pools, 0 bytes data, 0 objects >> 8460 MB used, 4162 GB / 4171 GB avail >> 64 creating >> >> >> How to deal with this HEALTH_WARN status? >> This is my ceph.conf: >> [global] >> >> >> fsid = 0b9b05db-98fe-49e6-b12b-1cce0645c015 >> >> >> mon initial members = monitorOne >> mon host = 192.168.1.153 >> filestore_xattr_use_omap = true >> >> >> public network = 192.168.1.0/24 >> cluster network = 10.0.0.0/24 >> pid file = /var/run/ceph/$name.pid >> >> >> auth cluster required = cephx >> auth service required = cephx >> auth client required = cephx >> >> >> osd pool default size = 3 >> osd pool default min size = 2 >> osd pool default pg num = 512 >> osd pool default pgp num = 512 >> osd crush chooseleaf type = 1 >> osd journal size = 1024 >> >> >> [mon] >> >> >> [mon.0] >> host = monitorOne >> mon addr = 192.168.1.153:6789 >> >> >> [osd] >> >> >> [osd.0] >> host = storageOne >> >> >> [osd.1] >> host = storageTwo >> >> >> [osd.2] >> host = storageFour >> >> >> [osd.3] >> host = storageLast >> >> >> Could anybody help me? >> >> best regards, >> >> -- >> >> Zhen Wang > >-- >Christian Balzer Network/Systems Engineer >chibi@xxxxxxx Global OnLine Japan/Fusion Communications >http://www.gol.com/


 

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux