On 02/09/2013 11:40 AM, Adam Nielsen wrote:
Hi again, I'm trying to mount my new Ceph volume on a remote PC using cephfs. I've followed the quick start guide, but when I try to mount the filesystem I get this: remote$ mount -t ceph 192.168.0.6:6789:/ /mnt/ceph/ mount: 192.168.0.6:6789:/: can't read superblock remote$ dmesg | tail [951382.981690] libceph: client4105 fsid aa447ff8-8270-491b-b59e-2735e852eaf5 [951382.983486] libceph: mon0 192.168.0.6:6789 session established I'm not sure what the problem is. The osd, mon and mds daemons are running on the Ceph host and no traffic is firewalled between the two machines. I have disabled authentication in ceph.conf. The only issue I can see is this: cephhost$ ceph health HEALTH_WARN 192 pgs degraded; 192 pgs stuck unclean
Could you show the output of "ceph -s" Also, which version of Ceph are you using under which OS? Wido
I'm not sure what that means as I have only used mkcephfs and then started the ceph service. I haven't actually used it yet so I'm not sure where the unclean pgs are coming from. The underlying filesystem is ext4 though. I'm a bit stuck as to what to do next. I can't see why the filesystem won't mount remotely, and the troubleshooting docs say unclean pgs are related to outages which doesn't seem possible on a newly created volume with only one osd/mon/mds instance. Any pointers would be much appreciated! Many thanks, Adam. -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html
-- Wido den Hollander 42on B.V. Phone: +31 (0)20 700 9902 Skype: contact42on -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html