Placement groups forever in "creating" state and dont map to OSD

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Dell - Internal Use - Confidential
Matt
Thanks for responding
As suggested I tried to set replication to 2X by usng commands you provided

$ceph osd pool set data size 2
$ceph osd pool set data min_size 2
$ceph osd pool set rbd size 2
$ceph osd pool set rbd min_size 2
$ceph osd pool set metadata size 2
$ceph osd pool set metadata min_size 2

It told me -
set pool 0 size to 2
set pool 0 min_size to 2
set pool 2 size to 2
set pool 2 min_size to 2
set pool 1 size to 2
set pool 1 min_size to 2

To verify that pool size had indeed changed - I checked again

$ceph osd dump | grep 'rep size'
pool 0 'data' rep size 2 crush_ruleset 0 object_hash rjenkins pg_num 64 pgp_num 64 last_change 90 owner 0 crash_replay_interval 45
pool 1 'metadata' rep size 2 crush_ruleset 1 object_hash rjenkins pg_num 64 pgp_num 64 last_change 94 owner 0
pool 2 'rbd' rep size 2 crush_ruleset 2 object_hash rjenkins pg_num 64 pgp_num 64 last_change 92 owner 0
pool 3 'datapool' rep size 2 crush_ruleset 2 object_hash rjenkins pg_num 10 pgp_num 10 last_change 38 owner 0


However - my cluster is still in same state

$ceph -s
   health HEALTH_WARN 202 pgs stuck inactive; 202 pgs stuck unclean
   monmap e1: 1 mons at {slesceph1=160.110.73.200:6789/0}, election epoch 1, quorum 0 slesceph1
   osdmap e106: 2 osds: 2 up, 2 in
    pgmap v171: 202 pgs: 202 creating; 0 bytes data, 10306 MB used, 71573 MB / 81880 MB avail
   mdsmap e1: 0/0/1 up
Yogesh Devi,
Architect,  Dell Cloud Clinical Archive
Dell


Land Phone     +91 80 28413000 Extension - 2781
Hand Phone    +91 99014 71082

From: Matt Harlum [mailto:matt@xxxxxxxxxxx]
Sent: Saturday, August 02, 2014 6:01 AM
To: Devi, Yogesh
Cc: Pulicken, Antony
Subject: Re: Placement groups forever in "creating" state and dont map to OSD

Hi Yogesh,

By default ceph is configured to create 3 replicas of the data, with only 3 OSDs it cannot create all of the pgs required to do this

You will need to change the replication to 2x for your pools, this can be done like so:
ceph odd pool set data size 2
ceph odd pool set data min_size 2
ceph odd pool set rbd size 2
ceph odd pool set rbd min_size 2
ceph odd pool set metadata size 2
ceph odd pool set metadata min_size 2

Once you do this your ceph cluster should go to a healthy state.

Regards,
Matt



On 2 Aug 2014, at 12:57 am, Yogesh_Devi at Dell.com<mailto:Yogesh_Devi at Dell.com> wrote:



Dell - Internal Use - Confidential
Hello Ceph Experts :) ,

I am using ceph ( ceph version 0.56.6) on Suse linux.
I created a simple cluster with one monitor server and two OSDs .
The conf file is attached

When  start my cluster - and do "ceph -s" -  I see following message

$ceph -s"
health HEALTH_WARN 202 pgs stuck inactive; 202 pgs stuck unclean
   monmap e1: 1 mons at {slesceph1=160.110.73.200:6789/0}, election epoch 1, quorum 0 slesceph1
   osdmap e56: 2 osds: 2 up, 2 in
    pgmap v100: 202 pgs: 202 creating; 0 bytes data, 10305 MB used, 71574 MB / 81880 MB avail
   mdsmap e1: 0/0/1 up


Basically there is some problem with my placement groups - they are forever stuck in "creating" state and there is no OSD associated with them ( despite having two OSD's that are up and in" ) - when I do a ceph pg stat" I see as follows

$ceph pg stat
v100: 202 pgs: 202 creating; 0 bytes data, 10305 MB used, 71574 MB / 81880 MB avail


if I query any individual pg - then I see it isn't mapped to any OSD
$ ceph pg 0.d query
pgid currently maps to no osd

I tried restaring OSDs and tuning my configuration without any avail

Any suggestions ?

Yogesh Devi
<ceph.conf>_______________________________________________
ceph-users mailing list
ceph-users at lists.ceph.com<mailto:ceph-users at lists.ceph.com>
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20140804/fdf8b65f/attachment.htm>


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux