Placement groups forever in "creating" state and dont map to OSD

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi

What distributions are your machines using? and is SELinux enabled on them?

I ran into the same issue once, i had to disable SELinux on all the machines and then reinstall


On 4 Aug 2014, at 5:25 pm, Yogesh_Devi at Dell.com wrote:

> Dell - Internal Use - Confidential
> 
> Matt
> Thanks for responding
> As suggested I tried to set replication to 2X by usng commands you provided
>  
> $ceph osd pool set data size 2
> $ceph osd pool set data min_size 2
> $ceph osd pool set rbd size 2
> $ceph osd pool set rbd min_size 2
> $ceph osd pool set metadata size 2
> $ceph osd pool set metadata min_size 2
>  
> It told me ?
> set pool 0 size to 2
> set pool 0 min_size to 2
> set pool 2 size to 2
> set pool 2 min_size to 2
> set pool 1 size to 2
> set pool 1 min_size to 2
>  
> To verify that pool size had indeed changed ? I checked again
>  
> $ceph osd dump | grep 'rep size'
> pool 0 'data' rep size 2 crush_ruleset 0 object_hash rjenkins pg_num 64 pgp_num 64 last_change 90 owner 0 crash_replay_interval 45
> pool 1 'metadata' rep size 2 crush_ruleset 1 object_hash rjenkins pg_num 64 pgp_num 64 last_change 94 owner 0
> pool 2 'rbd' rep size 2 crush_ruleset 2 object_hash rjenkins pg_num 64 pgp_num 64 last_change 92 owner 0
> pool 3 'datapool' rep size 2 crush_ruleset 2 object_hash rjenkins pg_num 10 pgp_num 10 last_change 38 owner 0
>  
>  
> However ? my cluster is still in same state
>  
> $ceph -s
>    health HEALTH_WARN 202 pgs stuck inactive; 202 pgs stuck unclean
>    monmap e1: 1 mons at {slesceph1=160.110.73.200:6789/0}, election epoch 1, quorum 0 slesceph1
>    osdmap e106: 2 osds: 2 up, 2 in
>     pgmap v171: 202 pgs: 202 creating; 0 bytes data, 10306 MB used, 71573 MB / 81880 MB avail
>    mdsmap e1: 0/0/1 up
> Yogesh Devi,
> Architect,  Dell Cloud Clinical Archive
> Dell 
>  
>  
> Land Phone     +91 80 28413000 Extension ? 2781
> Hand Phone    +91 99014 71082
>  
> From: Matt Harlum [mailto:matt at cactuar.net] 
> Sent: Saturday, August 02, 2014 6:01 AM
> To: Devi, Yogesh
> Cc: Pulicken, Antony
> Subject: Re: [ceph-users] Placement groups forever in "creating" state and dont map to OSD
>  
> Hi Yogesh,
>  
> By default ceph is configured to create 3 replicas of the data, with only 3 OSDs it cannot create all of the pgs required to do this
>  
> You will need to change the replication to 2x for your pools, this can be done like so:
> ceph odd pool set data size 2
> ceph odd pool set data min_size 2
> ceph odd pool set rbd size 2
> ceph odd pool set rbd min_size 2
> ceph odd pool set metadata size 2
> ceph odd pool set metadata min_size 2
>  
> Once you do this your ceph cluster should go to a healthy state.
>  
> Regards,
> Matt
>  
>  
>  
> On 2 Aug 2014, at 12:57 am, Yogesh_Devi at Dell.com wrote:
> 
> 
> Dell - Internal Use - Confidential
> 
> Hello Ceph Experts J ,
>  
> I am using ceph ( ceph version 0.56.6) on Suse linux.
> I created a simple cluster with one monitor server and two OSDs .
> The conf file is attached
>  
> When  start my cluster ? and do ?ceph ?s? -  I see following message
>  
> $ceph ?s?
> health HEALTH_WARN 202 pgs stuck inactive; 202 pgs stuck unclean
>    monmap e1: 1 mons at {slesceph1=160.110.73.200:6789/0}, election epoch 1, quorum 0 slesceph1
>    osdmap e56: 2 osds: 2 up, 2 in
>     pgmap v100: 202 pgs: 202 creating; 0 bytes data, 10305 MB used, 71574 MB / 81880 MB avail
>    mdsmap e1: 0/0/1 up
>  
>  
> Basically there is some problem with my placement groups ? they are forever stuck in ?creating? state and there is no OSD associated with them ( despite having two OSD?s that are up and in? ) ? when I do a ceph pg stat? I see as follows
>  
> $ceph pg stat
> v100: 202 pgs: 202 creating; 0 bytes data, 10305 MB used, 71574 MB / 81880 MB avail
>  
>  
> if I query any individual pg ? then I see it isn?t mapped to any OSD
> $ ceph pg 0.d query
> pgid currently maps to no osd
>  
> I tried restaring OSDs and tuning my configuration without any avail
>  
> Any suggestions ?
>  
> Yogesh Devi
> <ceph.conf>_______________________________________________
> ceph-users mailing list
> ceph-users at lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20140804/bf45d255/attachment.htm>


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux