Re: crush update failure ("corrupt inc osdmap epoch 5 off 50")

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Am Wed, 16 Jun 2010 09:57:56 -0700 schrieb Sage Weil:

> Hi Thomas,
> 
> On Wed, 16 Jun 2010, Thomas Mueller wrote:
>> i tried to change data "max_size" to 1 to disable replication for
>> testing.
> 
> You don't actually need/want to change the crush map to adjust
> replication.  Each crush rule is written to generalize to some range of
> replication levels (hence the min/max size definitions).  To adjust
> replication for testing/whatever, just do
> 
>  $ ceph osd pool data size 1   # 1x replication
> 
> See
> 
>  http://ceph.newdream.net/wiki/Adjusting_replication_level

ahh, once more i overlooked usefull information.

i've read http://ceph.newdream.net/wiki/Custom_data_placement_with_CRUSH 
and somehow came to the conclusion that this has to be done with an 
update of the crush map. 

the

>  $ ceph osd pool data size 1   # 1x replication

command worked. 

thanks!

- Thomas

--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux