Re: pgs stuck inactive and unclean, too feww PGs per OSD

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello,

On Thu, 8 Oct 2015 12:21:40 +0800 (CST) wikison wrote:

> Here, like this :
> esta@monitorOne:~$ sudo ceph osd tree
> ID WEIGHT  TYPE NAME            UP/DOWN REWEIGHT PRIMARY-AFFINITY
> -3 4.39996 root defualt
                 ^^^^^^^^
That's your problem. It should be "default" 

Your manually edited the crush map, right?

> -2 1.09999     host storageTwo
>  0 0.09999         osd.0             up  1.00000          1.00000
>  1 1.00000         osd.1             up  1.00000          1.00000
> -4 1.09999     host storageFour
>  2 0.09999         osd.2             up  1.00000          1.00000
>  3 1.00000         osd.3             up  1.00000          1.00000
> -5 1.09999     host storageLast
>  4 0.09999         osd.4             up  1.00000          1.00000
>  5 1.00000         osd.5             up  1.00000          1.00000
> -6 1.09999     host storageOne
>  6 0.09999         osd.6             up  1.00000          1.00000
>  7 1.00000         osd.7             up  1.00000          1.00000
> -1       0 root default
>
Nothing under the default root, so the default rule to allocate PGs can't
find anything.
 
Christian

> I have four storage nodes. Each of them has two independent hard drive
> to store data. One is 120GB SSD, and the other is 1TB HDD. I set the
> weight of SSD is 0.1 and weight of HDD is 1.0.
> 
> 
> 
> 
> 
> --
> 
> Zhen Wang
> Shanghai Jiao Tong University
> 
> 
> 
> At 2015-10-08 11:32:52, "Christian Balzer" <chibi@xxxxxxx> wrote:
> >
> >Hello,
> >
> >On Thu, 8 Oct 2015 11:27:46 +0800 (CST) wikison wrote:
> >
> >> Hi,
> >>         I've removed the rbd pool and created it again. It picked up
> >> my default settings but there are still some problems. After running
> >> "sudo ceph -s", the output is as follow: 
> >>     cluster 0b9b05db-98fe-49e6-b12b-1cce0645c015
> >>      health HEALTH_WARN
> >>             512 pgs stuck inactive
> >>             512 pgs stuck unclean
> >>      monmap e1: 1 mons at {monitorOne=192.168.1.153:6789/0}
> >>             election epoch 1, quorum 0 monitorOne
> >>      osdmap e62: 8 osds: 8 up, 8 in
> >>       pgmap v219: 512 pgs, 1 pools, 0 bytes data, 0 objects
> >>             8460 MB used, 4162 GB / 4171 GB avail
> >>                  512 creating
> >> 
> >Output of "ceph osd tree" please.
> >
> >The only reason I can think of is if your OSDs are up, but have no
> >weight.
> >
> >Christian
> >
> >> Ceph stucks in creating the pgs forever. Those pgs are stuck in
> >> inactive and unclean. And the Ceph pg query hangs forever. I googled
> >> this problem and didn't get a clue. Is there anything I missed?
> >> Any idea to help me?
> >> 
> >> 
> >> --
> >> 
> >> Zhen Wang
> >> 
> >> 
> >> 
> >> At 2015-10-07 13:05:51, "Christian Balzer" <chibi@xxxxxxx> wrote:
> >> >
> >> >Hello,
> >> >On Wed, 7 Oct 2015 12:57:58 +0800 (CST) wikison wrote:
> >> >
> >> >This is a very old bug, misfeature. 
> >> >And creeps up every week or so here, google is your friend.
> >> >
> >> >> Hi, 
> >> >> I have a cluster of one monitor and eight OSDs. These OSDs are
> >> >> running on four hosts(each host has two OSDs). When I set up
> >> >> everything and started Ceph, I got this: esta@monitorOne:~$ sudo
> >> >> ceph -s [sudo] password for esta: cluster
> >> >> 0b9b05db-98fe-49e6-b12b-1cce0645c015 health HEALTH_WARN
> >> >>             64 pgs stuck inactive
> >> >>             64 pgs stuck unclean
> >> >>             too few PGs per OSD (8 < min 30)
> >> >
> >> >Those 3 lines tell you pretty much all there is wrong.
> >> >You did (correctly) set the defaul pg and pgp nums to something
> >> >sensible (512) in your ceph.conf.
> >> >Unfortunately when creating the initial pool (rbd) it still ignores
> >> >those settings.
> >> >
> >> >You could try to increase those for your pool, which may or may not
> >> >work.
> >> >
> >> >The easier and faster way is to remove the rbd pool and create it
> >> >again. This should pick up your default settings.
> >> >
> >> >Christian
> >> >
> >> >>      monmap e1: 1 mons at {monitorOne=192.168.1.153:6789/0}
> >> >>             election epoch 1, quorum 0 monitorOne
> >> >>      osdmap e58: 8 osds: 8 up, 8 in
> >> >>       pgmap v191: 64 pgs, 1 pools, 0 bytes data, 0 objects
> >> >>             8460 MB used, 4162 GB / 4171 GB avail
> >> >>                   64 creating
> >> >> 
> >> >> 
> >> >> How to deal with this HEALTH_WARN status?
> >> >> This is my ceph.conf:
> >> >> [global]
> >> >> 
> >> >> 
> >> >>     fsid                        =
> >> >> 0b9b05db-98fe-49e6-b12b-1cce0645c015
> >> >> 
> >> >> 
> >> >>     mon initial members         = monitorOne
> >> >>     mon host                    = 192.168.1.153
> >> >>     filestore_xattr_use_omap    = true
> >> >> 
> >> >> 
> >> >>     public network              = 192.168.1.0/24
> >> >>     cluster network             = 10.0.0.0/24
> >> >>     pid file                    = /var/run/ceph/$name.pid
> >> >> 
> >> >> 
> >> >>     auth cluster required      = cephx
> >> >>     auth service required      = cephx
> >> >>     auth client required       = cephx
> >> >> 
> >> >> 
> >> >>     osd pool default size       = 3
> >> >>     osd pool default min size   = 2
> >> >>     osd pool default pg num     = 512
> >> >>     osd pool default pgp num    = 512
> >> >>     osd crush chooseleaf type   = 1
> >> >>     osd journal size            = 1024
> >> >> 
> >> >> 
> >> >> [mon]
> >> >> 
> >> >> 
> >> >> [mon.0]
> >> >>     host = monitorOne
> >> >>     mon addr = 192.168.1.153:6789
> >> >> 
> >> >> 
> >> >> [osd]
> >> >> 
> >> >> 
> >> >> [osd.0]
> >> >>     host = storageOne
> >> >> 
> >> >> 
> >> >> [osd.1]
> >> >>     host = storageTwo
> >> >> 
> >> >> 
> >> >> [osd.2]
> >> >>     host = storageFour
> >> >> 
> >> >> 
> >> >> [osd.3]
> >> >>     host = storageLast
> >> >>                         
> >> >> 
> >> >> Could anybody help me?
> >> >> 
> >> >> best regards,
> >> >> 
> >> >> --
> >> >> 
> >> >> Zhen Wang
> >> >
> >> >-- 
> >> >Christian Balzer        Network/Systems Engineer                
> >> >chibi@xxxxxxx   	Global OnLine Japan/Fusion Communications
> >> >http://www.gol.com/
> >
> >
> >-- 
> >Christian Balzer        Network/Systems Engineer                
> >chibi@xxxxxxx   	Global OnLine Japan/Fusion Communications
> >http://www.gol.com/


-- 
Christian Balzer        Network/Systems Engineer                
chibi@xxxxxxx   	Global OnLine Japan/Fusion Communications
http://www.gol.com/
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux