Re: Problems with active+remapped PGs in Ceph 0.55

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Josh,

that was the right answer ! Thank you ! :)

Norbert

On 12.12.2012 09:57, Josh Durgin wrote:
On 12/11/2012 11:48 PM, norbi wrote:
Hi Ceph-List,

i have set up a Ceph-Cluster with 3 OSDs, 3 Mons, 2 MDS over three
server.
Server 1 has 2 ODSs (osd0,osd2) and one MON/MDS and Server 2 has only
osd2 and one MON + MDS.
Server 3 has only the third MON-Service.

All Servers are running Ceph 0.55 and Kernel 3.6.9 with Centos 6 64bit.

ceph osd shows me in first state

ceph osd tree

# id    weight  type name       up/down reweight
-1      3       pool default
-3      3               rack unknownrack
-4      1                       host unknownhost
1       1                               osd.1   up      1
2       1                               osd.2   up      1
0       1                               osd.0   up      1

ceph health is ok !

now i have edit the crush map to the following

# id    weight  type name       up/down reweight
-1      3       pool default
-3      3               rack unknownrack
-4      1                       host testblade01
1       1                               osd.1   up      1
-2      1                       host unknownhost
2       1                               osd.2   up      1
0       1                               osd.0   up      1


now ceph is remaping some existing PGs and there are my problems began...

ceph is stopping remapping some PGs and the status is "current state
active+remapped" but nothing happens... after about 14h the PGs are in
the same state and ceph health is in status HEALTH_WARN. there are no
lost or unfound object, i can read all files in ceph without problems
and i can write into ceph storage.

how can i find the problem or force remapping of the PGs ? i have looked
into the source code, but i dont find a comman like "ceph pg force_remap
1.a4". "ceph pg PGNUMBER" shows me that there are no unfound objects.

any help ?

I'm guessing you're hitting the issue with small numbers of devices and
legacy crush tunables described here:

http://ceph.com/docs/master/rados/operations/crush-map/#impact-of-legacy-values


Updating to use the recommended new tunables as described on that page
should fix the problem.

You can verify that this is the problem by checking the output of
'ceph pg dump' - it will show the up set of osds for the remapped pgs
as only a single osd, and being remapped to an acting set including two
osds.

Josh
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux