Hello, On Thu, 8 Oct 2015 11:27:46 +0800 (CST) wikison wrote: > Hi, > I've removed the rbd pool and created it again. It picked up my > default settings but there are still some problems. After running "sudo > ceph -s", the output is as follow: > cluster 0b9b05db-98fe-49e6-b12b-1cce0645c015 > health HEALTH_WARN > 512 pgs stuck inactive > 512 pgs stuck unclean > monmap e1: 1 mons at {monitorOne=192.168.1.153:6789/0} > election epoch 1, quorum 0 monitorOne > osdmap e62: 8 osds: 8 up, 8 in > pgmap v219: 512 pgs, 1 pools, 0 bytes data, 0 objects > 8460 MB used, 4162 GB / 4171 GB avail > 512 creating > Output of "ceph osd tree" please. The only reason I can think of is if your OSDs are up, but have no weight. Christian > Ceph stucks in creating the pgs forever. Those pgs are stuck in inactive > and unclean. And the Ceph pg query hangs forever. I googled this problem > and didn't get a clue. Is there anything I missed? > Any idea to help me? > > > -- > > Zhen Wang > > > > At 2015-10-07 13:05:51, "Christian Balzer" <chibi@xxxxxxx> wrote: > > > >Hello, > >On Wed, 7 Oct 2015 12:57:58 +0800 (CST) wikison wrote: > > > >This is a very old bug, misfeature. > >And creeps up every week or so here, google is your friend. > > > >> Hi, > >> I have a cluster of one monitor and eight OSDs. These OSDs are running > >> on four hosts(each host has two OSDs). When I set up everything and > >> started Ceph, I got this: esta@monitorOne:~$ sudo ceph -s [sudo] > >> password for esta: cluster 0b9b05db-98fe-49e6-b12b-1cce0645c015 > >> health HEALTH_WARN > >> 64 pgs stuck inactive > >> 64 pgs stuck unclean > >> too few PGs per OSD (8 < min 30) > > > >Those 3 lines tell you pretty much all there is wrong. > >You did (correctly) set the defaul pg and pgp nums to something sensible > >(512) in your ceph.conf. > >Unfortunately when creating the initial pool (rbd) it still ignores > >those settings. > > > >You could try to increase those for your pool, which may or may not > >work. > > > >The easier and faster way is to remove the rbd pool and create it again. > >This should pick up your default settings. > > > >Christian > > > >> monmap e1: 1 mons at {monitorOne=192.168.1.153:6789/0} > >> election epoch 1, quorum 0 monitorOne > >> osdmap e58: 8 osds: 8 up, 8 in > >> pgmap v191: 64 pgs, 1 pools, 0 bytes data, 0 objects > >> 8460 MB used, 4162 GB / 4171 GB avail > >> 64 creating > >> > >> > >> How to deal with this HEALTH_WARN status? > >> This is my ceph.conf: > >> [global] > >> > >> > >> fsid = 0b9b05db-98fe-49e6-b12b-1cce0645c015 > >> > >> > >> mon initial members = monitorOne > >> mon host = 192.168.1.153 > >> filestore_xattr_use_omap = true > >> > >> > >> public network = 192.168.1.0/24 > >> cluster network = 10.0.0.0/24 > >> pid file = /var/run/ceph/$name.pid > >> > >> > >> auth cluster required = cephx > >> auth service required = cephx > >> auth client required = cephx > >> > >> > >> osd pool default size = 3 > >> osd pool default min size = 2 > >> osd pool default pg num = 512 > >> osd pool default pgp num = 512 > >> osd crush chooseleaf type = 1 > >> osd journal size = 1024 > >> > >> > >> [mon] > >> > >> > >> [mon.0] > >> host = monitorOne > >> mon addr = 192.168.1.153:6789 > >> > >> > >> [osd] > >> > >> > >> [osd.0] > >> host = storageOne > >> > >> > >> [osd.1] > >> host = storageTwo > >> > >> > >> [osd.2] > >> host = storageFour > >> > >> > >> [osd.3] > >> host = storageLast > >> > >> > >> Could anybody help me? > >> > >> best regards, > >> > >> -- > >> > >> Zhen Wang > > > >-- > >Christian Balzer Network/Systems Engineer > >chibi@xxxxxxx Global OnLine Japan/Fusion Communications > >http://www.gol.com/ -- Christian Balzer Network/Systems Engineer chibi@xxxxxxx Global OnLine Japan/Fusion Communications http://www.gol.com/ _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com