Hi, It seems that ext4 and btrfs are not to be considered as stable for now. Does anyone could confirm that ext3 is the best choice for this moment ? On 17 October 2011 13:49, Martin Mailand <martin@xxxxxxxxxxxx> wrote: > Am 17.10.2011 11:40, schrieb Christian Brunner: >> >> 2011/10/15 Martin Mailand<martin@xxxxxxxxxxxx>: >>> >>> Hi Christian, >>> I have a very similar experience, I also used josef's tree and btrfs >>> snaps = >>> 0, the next problem I had than was excessive fragmentation, so I used >>> this >>> patch http://marc.info/?l=linux-btrfs&m=131495014823121&w=2, and changed >>> the >>> btrfs option to (btrfs options = noatime,nodatacow,autodefrag) that kept >>> the >>> fragmentation under control. >>> But even with this setup after a few days the load on the osd is >>> unbearable. >> >> How did you find out about our fragmentation issues? Was it just a >> performance problem? >> > > I used filefrag to show the number of extents, after the patch, I have on > average 1,14 extents per 4MB ceph object on the osd. > >>> As far as I understood the doku if you disable the btrfs snapshot >>> functionality the writeahead journal is activated. >>> http://ceph.newdream.net/wiki/Ceph.conf >>> And I get this in the logs. >>> mount: enabling WRITEAHEAD journal mode: 'filestore btrfs snap' mode is >>> not >>> enabled >>> >>> May I asked what kind of probs you did have with ext4? Because I am >>> looking >>> into this direction as well. >> >> You can read about our ext4 problems here: >> >> http://marc.info/?l=ceph-devel&m=131201869703245&w=2 > > I still can reproduce the bug with v3.1-rc9. > >> >> Our bugreport with RedHat didn't make any progress for a long time, >> but last week RedHat made two sugestions: >> >> - If you configure ceph with 'filestore flusher = false', do you see >> any different behavior? >> - If you mount with -o noauto_da_alloc does it change anything? >> >> Since I have just migrated to btrfs, I've some problems to check this, >> but I'll try to do this as soon as I can get hold of some extra >> hardware. >> > I can check this, I have a spare cluster at the moment. > >> Regards, >> Christian > > -- > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html > -- Tomasz Paszkowski SS7, Asterisk, SAN, Datacenter, Cloud Computing +48500166299 -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html