Placement groups forever in "creating" state and dont map to OSD

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I think ceph osd tree should list your OSDs under the node bucket.
Could you check your osd crush map also with this - 

ceph osd getcrushmap -o filename
crushtool -d filename -o filename.txt

you should see your OSDs in the #devices section and you
should see your three servers in the #buckets section.



Regards,
Kapil.




On Mon, 2014-08-04 at 10:09 +0000, Yogesh_Devi at Dell.com wrote:
> Dell - Internal Use - Confidential
> 
> Hi Kapil
> 
> Thanks for responding J 
> 
> My Mon-server two OSD?s are running on three separate servers one for
> respective node. All are SLES sp3.
> 
>  
> 
> Below is ?ceph osd tree? output from my mon server box 
> 
>  
> 
> slesceph1: # ceph osd tree
> 
> # id    weight  type name       up/down reweight
> 
> -1      1       root default
> 
> 0       0.5             osd.0   up      1
> 
> 1       0.5             osd.1   up      1
> 
> Yogesh  
> 
>  
> 
> Land Phone     +91 80 28413000 Extension ? 2781 
> Hand Phone    +91 99014 71082 
> 
>  
> 
> -----Original Message----- 
> From: Kapil Sharma [mailto:ksharma at suse.com]
> Sent: Monday, August 04, 2014 3:31 PM 
> To: Devi, Yogesh 
> Cc: matt at cactuar.net; ceph-users at lists.ceph.com; Pulicken, Antony 
> Subject: Re: [ceph-users] Placement groups forever in "creating" state
> and dont map to OSD 
> 
> Hi Yogesh, 
> 
> Are your two OSDs on same node ? Could you check the osd tree output
> with the command - "ceph osd tree" 
> 
>  
> 
> Regards, 
> Kapil. 
> 
> 
> 
> On Mon, 2014-08-04 at 09:22 +0000, Yogesh_Devi at Dell.com wrote: 
> > Dell - Internal Use - Confidential 
> > 
> > Matt 
> > 
> > I am using Suse Enterprise Linux 11 ? SP3 ( SLES SP3) 
> > 
> >  
> > 
> > I don?t think I have enabled SE Linux ..
> > 
> >  
> > 
> > Yogesh Devi, 
> > 
> > Architect,  Dell Cloud Clinical Archive 
> > 
> > Dell 
> > 
> >  
> > 
> >  
> > 
> > Land Phone     +91 80 28413000 Extension ? 2781 
> > 
> > Hand Phone    +91 99014 71082 
> > 
> > 
> >  
> > 
> > From: Matt Harlum [mailto:matt at cactuar.net] 
> > Sent: Monday, August 04, 2014 1:43 PM 
> > To: Devi, Yogesh 
> > Cc: ceph-users at lists.ceph.com; Pulicken, Antony 
> > Subject: Re: [ceph-users] Placement groups forever in "creating"
> state
> > and dont map to OSD 
> > 
> > 
> >  
> > 
> > Hi 
> > 
> >  
> > 
> > 
> > What distributions are your machines using? and is SELinux enabled
> on
> > them? 
> > 
> > 
> >  
> > 
> > 
> > I ran into the same issue once, i had to disable SELinux on all the
> > machines and then reinstall 
> > 
> > 
> >  
> > 
> > 
> >  
> > 
> > On 4 Aug 2014, at 5:25 pm, Yogesh_Devi at Dell.com wrote: 
> > 
> > 
> > 
> > 
> > Dell - Internal Use - Confidential 
> > 
> > Matt 
> > 
> > 
> > Thanks for responding 
> > 
> > 
> > As suggested I tried to set replication to 2X by usng commands you
> > provided 
> > 
> > 
> >  
> > 
> > 
> > $ceph osd pool set data size 2 
> > 
> > 
> > $ceph osd pool set data min_size 2 
> > 
> > 
> > $ceph osd pool set rbd size 2 
> > 
> > 
> > $ceph osd pool set rbd min_size 2 
> > 
> > 
> > $ceph osd pool set metadata size 2 
> > 
> > 
> > $ceph osd pool set metadata min_size 2 
> > 
> > 
> >  
> > 
> > 
> > It told me ? 
> > 
> > 
> > set pool 0 size to 2 
> > 
> > 
> > set pool 0 min_size to 2 
> > 
> > 
> > set pool 2 size to 2 
> > 
> > 
> > set pool 2 min_size to 2 
> > 
> > 
> > set pool 1 size to 2 
> > 
> > 
> > set pool 1 min_size to 2 
> > 
> > 
> >  
> > 
> > 
> > To verify that pool size had indeed changed ? I checked again 
> > 
> > 
> >  
> > 
> > 
> > $ceph osd dump | grep 'rep size' 
> > 
> > 
> > pool 0 'data' rep size 2 crush_ruleset 0 object_hash rjenkins
> pg_num 
> > 64 pgp_num 64 last_change 90 owner 0 crash_replay_interval 45 
> > 
> > 
> > pool 1 'metadata' rep size 2 crush_ruleset 1 object_hash rjenkins
> > pg_num 64 pgp_num 64 last_change 94 owner 0 
> > 
> > 
> > pool 2 'rbd' rep size 2 crush_ruleset 2 object_hash rjenkins pg_num
> 64
> > pgp_num 64 last_change 92 owner 0 
> > 
> > 
> > pool 3 'datapool' rep size 2 crush_ruleset 2 object_hash rjenkins
> > pg_num 10 pgp_num 10 last_change 38 owner 0 
> > 
> > 
> >  
> > 
> > 
> >  
> > 
> > 
> > However ? my cluster is still in same state 
> > 
> > 
> >  
> > 
> > 
> > $ceph -s 
> > 
> > 
> >    health HEALTH_WARN 202 pgs stuck inactive; 202 pgs stuck unclean 
> > 
> > 
> >    monmap e1: 1 mons at {slesceph1=160.110.73.200:6789/0}, election
> > epoch 1, quorum 0 slesceph1 
> > 
> > 
> >    osdmap e106: 2 osds: 2 up, 2 in 
> > 
> > 
> >     pgmap v171: 202 pgs: 202 creating; 0 bytes data, 10306 MB used, 
> > 71573 MB / 81880 MB avail 
> > 
> > 
> >    mdsmap e1: 0/0/1 up 
> > 
> > 
> > Yogesh Devi, 
> > 
> > 
> > Architect,  Dell Cloud Clinical Archive 
> > 
> > 
> > Dell 
> > 
> > 
> >  
> > 
> > 
> >  
> > 
> > 
> > Land Phone     +91 80 28413000 Extension ? 2781 
> > 
> > 
> > Hand Phone    +91 99014 71082 
> > 
> > 
> >  
> > 
> > 
> > From: Matt Harlum [mailto:matt at cactuar.net] 
> > Sent: Saturday, August 02, 2014 6:01 AM 
> > To: Devi, Yogesh 
> > Cc: Pulicken, Antony 
> > Subject: Re: [ceph-users] Placement groups forever in "creating"
> state
> > and dont map to OSD 
> > 
> > 
> >  
> > 
> > 
> > Hi Yogesh, 
> > 
> > 
> >  
> > 
> > 
> > By default ceph is configured to create 3 replicas of the data, with
> > only 3 OSDs it cannot create all of the pgs required to do this 
> > 
> > 
> >  
> > 
> > 
> > You will need to change the replication to 2x for your pools, this
> can
> > be done like so: 
> > 
> > 
> > ceph odd pool set data size 2 
> > 
> > 
> > ceph odd pool set data min_size 2 
> > 
> > 
> > ceph odd pool set rbd size 2 
> > 
> > 
> > ceph odd pool set rbd min_size 2 
> > 
> > 
> > ceph odd pool set metadata size 2 
> > 
> > 
> > ceph odd pool set metadata min_size 2 
> > 
> > 
> >  
> > 
> > 
> > Once you do this your ceph cluster should go to a healthy state. 
> > 
> > 
> >  
> > 
> > 
> > Regards, 
> > 
> > 
> > Matt 
> > 
> > 
> >  
> > 
> > 
> >  
> > 
> > 
> >  
> > 
> > 
> > On 2 Aug 2014, at 12:57 am, Yogesh_Devi at Dell.com wrote: 
> > 
> > 
> > 
> > 
> > 
> > Dell - Internal Use - Confidential 
> > 
> > Hello Ceph Experts J , 
> > 
> > 
> >  
> > 
> > 
> > I am using ceph ( ceph version 0.56.6) on Suse linux. 
> > 
> > 
> > I created a simple cluster with one monitor server and two OSDs . 
> > 
> > 
> > The conf file is attached 
> > 
> > 
> >  
> > 
> > 
> > When  start my cluster ? and do ?ceph ?s? -  I see following
> message 
> > 
> > 
> >  
> > 
> > 
> > $ceph ?s? 
> > 
> > 
> > health HEALTH_WARN 202 pgs stuck inactive; 202 pgs stuck unclean 
> > 
> > 
> >    monmap e1: 1 mons at {slesceph1=160.110.73.200:6789/0}, election
> > epoch 1, quorum 0 slesceph1 
> > 
> > 
> >    osdmap e56: 2 osds: 2 up, 2 in 
> > 
> > 
> >     pgmap v100: 202 pgs: 202 creating; 0 bytes data, 10305 MB used, 
> > 71574 MB / 81880 MB avail 
> > 
> > 
> >    mdsmap e1: 0/0/1 up 
> > 
> > 
> >  
> > 
> > 
> >  
> > 
> > 
> > Basically there is some problem with my placement groups ? they are
> > forever stuck in ?creating? state and there is no OSD associated
> with
> > them ( despite having two OSD?s that are up and in? ) ? when I do a
> > ceph pg stat? I see as follows 
> > 
> > 
> >  
> > 
> > 
> > $ceph pg stat 
> > 
> > 
> > v100: 202 pgs: 202 creating; 0 bytes data, 10305 MB used, 71574
> MB / 
> > 81880 MB avail 
> > 
> > 
> >  
> > 
> > 
> >  
> > 
> > 
> > if I query any individual pg ? then I see it isn?t mapped to any
> OSD 
> > 
> > 
> > $ ceph pg 0.d query 
> > 
> > 
> > pgid currently maps to no osd 
> > 
> > 
> >  
> > 
> > 
> > I tried restaring OSDs and tuning my configuration without any
> avail 
> > 
> > 
> >  
> > 
> > 
> > Any suggestions ? 
> > 
> > 
> >  
> > 
> > 
> > Yogesh Devi 
> > 
> > 
> > <ceph.conf>_______________________________________________ 
> > ceph-users mailing list 
> > ceph-users at lists.ceph.com 
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
> > 
> > 
> >  
> > 
> > 
> > _______________________________________________ 
> > ceph-users mailing list 
> > ceph-users at lists.ceph.com 
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
> 
>  
> 
> 




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux