On Fri, Sep 27, 2013 at 2:44 PM, Gregory Farnum <greg@xxxxxxxxxxx> wrote:
What is the output of ceph -s? It could be something underneath the filesystem.
root@chekov:~# ceph -s
cluster 18b7cba7-ccc3-4945-bb39-99450be81c98
health HEALTH_OK
monmap e3: 3 mons at {chekov=10.42.6.29:6789/0,laforge=10.42.5.30:6789/0,picard=10.42.6.21:6789/0}, election epoch 30, quorum 0,1,2 chekov,laforge,picard
osdmap e387: 4 osds: 4 up, 4 in
pgmap v1100: 320 pgs: 320 active+clean; 7568 MB data, 15445 MB used, 14873 GB / 14888 GB avail
mdsmap e28: 1/1/1 up {0=1=up:active}, 2 up:standby
cluster 18b7cba7-ccc3-4945-bb39-99450be81c98
health HEALTH_OK
monmap e3: 3 mons at {chekov=10.42.6.29:6789/0,laforge=10.42.5.30:6789/0,picard=10.42.6.21:6789/0}, election epoch 30, quorum 0,1,2 chekov,laforge,picard
osdmap e387: 4 osds: 4 up, 4 in
pgmap v1100: 320 pgs: 320 active+clean; 7568 MB data, 15445 MB used, 14873 GB / 14888 GB avail
mdsmap e28: 1/1/1 up {0=1=up:active}, 2 up:standby
What kernel version are you using?
I have two configurations I've tested:
root@chekov:~# uname -a
Linux chekov 3.5.0-40-generic #62~precise1-Ubuntu SMP Fri Aug 23 17:38:26 UTC 2013 x86_64 x86_64 x86_64 GNU/Linux
aaron@seven ~ $ uname -a
Linux seven 3.10.7-gentoo-r1 #1 SMP PREEMPT Thu Sep 26 07:23:03 PDT 2013 x86_64 Intel(R) Core(TM) i7-3770K CPU @ 3.50GHz GenuineIntel GNU/Linux
Linux chekov 3.5.0-40-generic #62~precise1-Ubuntu SMP Fri Aug 23 17:38:26 UTC 2013 x86_64 x86_64 x86_64 GNU/Linux
aaron@seven ~ $ uname -a
Linux seven 3.10.7-gentoo-r1 #1 SMP PREEMPT Thu Sep 26 07:23:03 PDT 2013 x86_64 Intel(R) Core(TM) i7-3770K CPU @ 3.50GHz GenuineIntel GNU/Linux
Did you enable crush tunables?
I haven't edited the CRUSH map for this cluster after creating it... so I don't think I've edited any of the tunables either.
It
could be that your kernel doesn't support all the options you enabled.
Well, on both my Ubuntu and Gentoo systems, I can mount the other cluster just fine (the one that started as 0.61 and got upgraded):
seven ~ # mount -t ceph 10.42.100.20:/ /mnt/ceph -o name=admin,secret=...
seven ~ # mount -t ceph 10.42.100.20:/ /mnt/ceph -o name=admin,secret=...
seven ~ #
I should have mentioned in my initial email that with or without -o name=,secret= mounting the new cluster fails with the same error 95 = Operation not supported.
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com