On Thu, May 03, 2012 at 08:17:43AM -0700, Josh Durgin wrote: > On Thu, 3 May 2012 10:13:55 -0400, Josef Bacik <josef@xxxxxxxxxx> > wrote: > > On Fri, Apr 27, 2012 at 01:02:08PM +0200, Christian Brunner wrote: > >> Am 24. April 2012 18:26 schrieb Sage Weil <sage@xxxxxxxxxxxx>: > >> > On Tue, 24 Apr 2012, Josef Bacik wrote: > >> >> On Fri, Apr 20, 2012 at 05:09:34PM +0200, Christian Brunner wrote: > >> >> > After running ceph on XFS for some time, I decided to try btrfs again. > >> >> > Performance with the current "for-linux-min" branch and big metadata > >> >> > is much better. The only problem (?) I'm still seeing is a warning > >> >> > that seems to occur from time to time: > >> > > >> > Actually, before you do that... we have a new tool, > >> > test_filestore_workloadgen, that generates a ceph-osd-like workload on the > >> > local file system. It's a subset of what a full OSD might do, but if > >> > we're lucky it will be sufficient to reproduce this issue. Something like > >> > > >> > test_filestore_workloadgen --osd-data /foo --osd-journal /bar > >> > > >> > will hopefully do the trick. > >> > > >> > Christian, maybe you can see if that is able to trigger this warning? > >> > You'll need to pull it from the current master branch; it wasn't in the > >> > last release. > >> > >> Trying to reproduce with test_filestore_workloadgen didn't work for > >> me. So here are some instructions on how to reproduce with a minimal > >> ceph setup. > >> > >> You will need a single system with two disks and a bit of memory. > >> > >> - Compile and install ceph (detailed instructions: > >> http://ceph.newdream.net/docs/master/ops/install/mkcephfs/) > >> > >> - For the test setup I've used two tmpfs files as journal devices. To > >> create these, do the following: > >> > >> # mkdir -p /ceph/temp > >> # mount -t tmpfs tmpfs /ceph/temp > >> # dd if=/dev/zero of=/ceph/temp/journal0 count=500 bs=1024k > >> # dd if=/dev/zero of=/ceph/temp/journal1 count=500 bs=1024k > >> > >> - Now you should create and mount btrfs. Here is what I did: > >> > >> # mkfs.btrfs -l 64k -n 64k /dev/sda > >> # mkfs.btrfs -l 64k -n 64k /dev/sdb > >> # mkdir /ceph/osd.000 > >> # mkdir /ceph/osd.001 > >> # mount -o noatime,space_cache,inode_cache,autodefrag /dev/sda /ceph/osd.000 > >> # mount -o noatime,space_cache,inode_cache,autodefrag /dev/sdb /ceph/osd.001 > >> > >> - Create /etc/ceph/ceph.conf similar to the attached ceph.conf. You > >> will probably have to change the btrfs devices and the hostname > >> (os39). > >> > >> - Create the ceph filesystems: > >> > >> # mkdir /ceph/mon > >> # mkcephfs -a -c /etc/ceph/ceph.conf > >> > >> - Start ceph (e.g. "service ceph start") > >> > >> - Now you should be able to use ceph - "ceph -s" will tell you about > >> the state of the ceph cluster. > >> > >> - "rbd create -size 100 testimg" will create an rbd image on the ceph cluster. > >> > > > > It's failing here > > > > http://fpaste.org/e3BG/ > > 2012-05-03 10:11:28.818308 7fcb5a0ee700 -- 127.0.0.1:0/1003269 <== > osd.1 127.0.0.1:6803/2379 3 ==== osd_op_reply(3 rbd_info [call] = -5 > (Input/output error)) v4 ==== 107+0+0 (3948821281 0 0) 0x7fcb380009a0 > con 0x1cad3e0 > > This is probably because the osd isn't finding the rbd class. > Do you have 'rbd_cls.so' in /usr/lib64/rados-classes? Wherever > rbd_cls.so is, > try adding 'osd class dir = /path/to/rados-classes' to the [osd] > section > in your ceph.conf, and restarting the osds. > > If you set 'debug osd = 10' you should see '_load_class rbd' in the osd > log > when you try to create an rbd image. > > Autotools should be setting the default location correctly, but if > you're > running the osds in a chroot or something the path would be wrong. > Yeah all that was in the right place, I rebooted and I magically stopped getting that error, but now I'm getting this http://fpaste.org/OE92/ with that ping thing repeating over and over. Thanks, Josef -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html