Hey all, We are currently testing CephFS on a small (3 node) cluster. The setup is currently: Each server has 12 OSDs, 1 Monitor and 1 MDS running on it: The servers are running: 0.94.2-0.el7 The clients are running: Ceph: 0.80.10-1.fc21, Kernel: 4.0.6-200.fc21.x86_64 ceph -s cluster 4ed5ecdd-0c5b-4422-9d99-c9e42c6bd4cd health HEALTH_OK monmap e1: 3 mons at {ceph1=10.15.0.1:6789/0,ceph2=10.15.0.2:6789/0,ceph3=10.15.0.3:6789/0} election epoch 20, quorum 0,1,2 ceph1,ceph2,ceph3 mdsmap e12: 1/1/1 up {0=ceph3=up:active}, 2 up:standby osdmap e389: 36 osds: 36 up, 36 in pgmap v19370: 8256 pgs, 3 pools, 51217 MB data, 14035 objects 95526 MB used, 196 TB / 196 TB avail 8256 active+clean Our Ceph.conf is relatively simple at the moment: cat /etc/ceph/ceph.conf [global] fsid = 4ed5ecdd-0c5b-4422-9d99-c9e42c6bd4cd mon_initial_members = ceph1, ceph2, ceph3 mon_host = 10.15.0.1,10.15.0.2,10.15.0.3 mon_pg_warn_max_per_osd = 1000 auth_cluster_required = cephx auth_service_required = cephx auth_client_required = cephx filestore_xattr_use_omap = true osd_pool_default_size = 2 When I pulled the plug on the master MDS last time (ceph1), it stopped all IO until I plugged it back in. I was under the assumption that the MDS would fail over the other 2 MDS’s and IO would continue? Is there something I need to do to allow the MDS’s to failover from each other without too much interruption? Or is this because the clients ceph version? Cheers, Simon Hallam Linux Support & Development Officer Please visit our new website at www.pml.ac.uk and follow us on Twitter @PlymouthMarine Winner of the Environment & Conservation category, the Charity Awards 2014. Plymouth Marine Laboratory (PML) is a company limited by guarantee registered in England & Wales, company number 4178503. Registered Charity No. 1091222. Registered Office: Prospect Place, The Hoe, Plymouth PL1 3DH, UK. This message is private and confidential. If you have received this message in error, please notify the sender and remove it from your system. You are reminded that e-mail communications are not secure and may contain viruses; PML accepts no liability for any loss or damage which may be caused by viruses. |
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com