I have two test clusters running Bobtail (0.56.4) and Ubuntu Precise (12.04.2). The problem I'm having is that I'm not able to get either of them into a state where I can both mount the filesystem and have all the PGs in the active+clean state. It seems that on both clusters I can get them into a 100% active+clean state by setting "ceph osd crush tunables bobtail", but when I try to mount the filesystem I get: mount error 5 = Input/output error However, if I set "ceph osd crush tunables legacy" I can mount both filesystems, but then some of the PGs are stuck in the "active+remapped" state: # ceph -s health HEALTH_WARN 29 pgs stuck unclean; recovery 5/1604152 degraded (0.000%) monmap e1: 1 mons at {a=172.16.0.50:6789/0}, election epoch 1, quorum 0 a osdmap e10272: 20 osds: 20 up, 20 in pgmap v1114740: 1920 pgs: 1890 active+clean, 29 active+remapped, 1 active+clean+scrubbing; 3086 GB data, 6201 GB used, 3098 GB / 9300 GB avail; 232B/s wr, 0op/s; 5/1604152 degraded (0.000%) mdsmap e420: 1/1/1 up {0=a=up:active} Is any one else seeing this? Thanks, Bryan _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com