Placement groups forever in "creating" state and dont map to OSD

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Yogesh, 

Are your two OSDs on same node ? Could you check the osd tree output
with the command - "ceph osd tree"


Regards,
Kapil.




On Mon, 2014-08-04 at 09:22 +0000, Yogesh_Devi at Dell.com wrote:
> Dell - Internal Use - Confidential
> 
> Matt
> 
> I am using Suse Enterprise Linux 11 ? SP3 ( SLES SP3) 
> 
>  
> 
> I don?t think I have enabled SE Linux .. 
> 
>  
> 
> Yogesh Devi,
> 
> Architect,  Dell Cloud Clinical Archive
> 
> Dell 
> 
>  
> 
>  
> 
> Land Phone     +91 80 28413000 Extension ? 2781
> 
> Hand Phone    +91 99014 71082 
> 
> 
>  
> 
> From: Matt Harlum [mailto:matt at cactuar.net] 
> Sent: Monday, August 04, 2014 1:43 PM
> To: Devi, Yogesh
> Cc: ceph-users at lists.ceph.com; Pulicken, Antony
> Subject: Re: [ceph-users] Placement groups forever in "creating" state
> and dont map to OSD
> 
> 
>  
> 
> Hi
> 
>  
> 
> 
> What distributions are your machines using? and is SELinux enabled on
> them?
> 
> 
>  
> 
> 
> I ran into the same issue once, i had to disable SELinux on all the
> machines and then reinstall
> 
> 
>  
> 
> 
>  
> 
> On 4 Aug 2014, at 5:25 pm, Yogesh_Devi at Dell.com wrote:
> 
> 
> 
> 
> Dell - Internal Use - Confidential
> 
> Matt
> 
> 
> Thanks for responding
> 
> 
> As suggested I tried to set replication to 2X by usng commands you
> provided
> 
> 
>  
> 
> 
> $ceph osd pool set data size 2
> 
> 
> $ceph osd pool set data min_size 2
> 
> 
> $ceph osd pool set rbd size 2
> 
> 
> $ceph osd pool set rbd min_size 2
> 
> 
> $ceph osd pool set metadata size 2
> 
> 
> $ceph osd pool set metadata min_size 2
> 
> 
>  
> 
> 
> It told me ?
> 
> 
> set pool 0 size to 2
> 
> 
> set pool 0 min_size to 2
> 
> 
> set pool 2 size to 2
> 
> 
> set pool 2 min_size to 2
> 
> 
> set pool 1 size to 2
> 
> 
> set pool 1 min_size to 2
> 
> 
>  
> 
> 
> To verify that pool size had indeed changed ? I checked again
> 
> 
>  
> 
> 
> $ceph osd dump | grep 'rep size'
> 
> 
> pool 0 'data' rep size 2 crush_ruleset 0 object_hash rjenkins pg_num
> 64 pgp_num 64 last_change 90 owner 0 crash_replay_interval 45
> 
> 
> pool 1 'metadata' rep size 2 crush_ruleset 1 object_hash rjenkins
> pg_num 64 pgp_num 64 last_change 94 owner 0
> 
> 
> pool 2 'rbd' rep size 2 crush_ruleset 2 object_hash rjenkins pg_num 64
> pgp_num 64 last_change 92 owner 0
> 
> 
> pool 3 'datapool' rep size 2 crush_ruleset 2 object_hash rjenkins
> pg_num 10 pgp_num 10 last_change 38 owner 0
> 
> 
>  
> 
> 
>  
> 
> 
> However ? my cluster is still in same state
> 
> 
>  
> 
> 
> $ceph -s
> 
> 
>    health HEALTH_WARN 202 pgs stuck inactive; 202 pgs stuck unclean
> 
> 
>    monmap e1: 1 mons at {slesceph1=160.110.73.200:6789/0}, election
> epoch 1, quorum 0 slesceph1
> 
> 
>    osdmap e106: 2 osds: 2 up, 2 in
> 
> 
>     pgmap v171: 202 pgs: 202 creating; 0 bytes data, 10306 MB used,
> 71573 MB / 81880 MB avail
> 
> 
>    mdsmap e1: 0/0/1 up
> 
> 
> Yogesh Devi,
> 
> 
> Architect,  Dell Cloud Clinical Archive
> 
> 
> Dell 
> 
> 
>  
> 
> 
>  
> 
> 
> Land Phone     +91 80 28413000 Extension ? 2781
> 
> 
> Hand Phone    +91 99014 71082
> 
> 
>  
> 
> 
> From: Matt Harlum [mailto:matt at cactuar.net] 
> Sent: Saturday, August 02, 2014 6:01 AM
> To: Devi, Yogesh
> Cc: Pulicken, Antony
> Subject: Re: [ceph-users] Placement groups forever in "creating" state
> and dont map to OSD
> 
> 
>  
> 
> 
> Hi Yogesh,
> 
> 
>  
> 
> 
> By default ceph is configured to create 3 replicas of the data, with
> only 3 OSDs it cannot create all of the pgs required to do this
> 
> 
>  
> 
> 
> You will need to change the replication to 2x for your pools, this can
> be done like so:
> 
> 
> ceph odd pool set data size 2
> 
> 
> ceph odd pool set data min_size 2
> 
> 
> ceph odd pool set rbd size 2
> 
> 
> ceph odd pool set rbd min_size 2
> 
> 
> ceph odd pool set metadata size 2
> 
> 
> ceph odd pool set metadata min_size 2
> 
> 
>  
> 
> 
> Once you do this your ceph cluster should go to a healthy state.
> 
> 
>  
> 
> 
> Regards,
> 
> 
> Matt
> 
> 
>  
> 
> 
>  
> 
> 
>  
> 
> 
> On 2 Aug 2014, at 12:57 am, Yogesh_Devi at Dell.com wrote:
> 
> 
> 
> 
> 
> Dell - Internal Use - Confidential
> 
> Hello Ceph Experts J ,
> 
> 
>  
> 
> 
> I am using ceph ( ceph version 0.56.6) on Suse linux.
> 
> 
> I created a simple cluster with one monitor server and two OSDs .
> 
> 
> The conf file is attached
> 
> 
>  
> 
> 
> When  start my cluster ? and do ?ceph ?s? -  I see following message
> 
> 
>  
> 
> 
> $ceph ?s?
> 
> 
> health HEALTH_WARN 202 pgs stuck inactive; 202 pgs stuck unclean
> 
> 
>    monmap e1: 1 mons at {slesceph1=160.110.73.200:6789/0}, election
> epoch 1, quorum 0 slesceph1
> 
> 
>    osdmap e56: 2 osds: 2 up, 2 in
> 
> 
>     pgmap v100: 202 pgs: 202 creating; 0 bytes data, 10305 MB used,
> 71574 MB / 81880 MB avail
> 
> 
>    mdsmap e1: 0/0/1 up
> 
> 
>  
> 
> 
>  
> 
> 
> Basically there is some problem with my placement groups ? they are
> forever stuck in ?creating? state and there is no OSD associated with
> them ( despite having two OSD?s that are up and in? ) ? when I do a
> ceph pg stat? I see as follows
> 
> 
>  
> 
> 
> $ceph pg stat
> 
> 
> v100: 202 pgs: 202 creating; 0 bytes data, 10305 MB used, 71574 MB /
> 81880 MB avail
> 
> 
>  
> 
> 
>  
> 
> 
> if I query any individual pg ? then I see it isn?t mapped to any OSD
> 
> 
> $ ceph pg 0.d query
> 
> 
> pgid currently maps to no osd
> 
> 
>  
> 
> 
> I tried restaring OSDs and tuning my configuration without any avail
> 
> 
>  
> 
> 
> Any suggestions ?
> 
> 
>  
> 
> 
> Yogesh Devi
> 
> 
> <ceph.conf>_______________________________________________
> ceph-users mailing list
> ceph-users at lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 
> 
>  
> 
> 
> _______________________________________________
> ceph-users mailing list
> ceph-users at lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux