Testing CephFS

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hey all,

 

We are currently testing CephFS on a small (3 node) cluster.

 

The setup is currently:

 

Each server has 12 OSDs, 1 Monitor and 1 MDS running on it:

The servers are running: 0.94.2-0.el7

The clients are running: Ceph: 0.80.10-1.fc21, Kernel: 4.0.6-200.fc21.x86_64

 

ceph -s

    cluster 4ed5ecdd-0c5b-4422-9d99-c9e42c6bd4cd

     health HEALTH_OK

     monmap e1: 3 mons at {ceph1=10.15.0.1:6789/0,ceph2=10.15.0.2:6789/0,ceph3=10.15.0.3:6789/0}

            election epoch 20, quorum 0,1,2 ceph1,ceph2,ceph3

     mdsmap e12: 1/1/1 up {0=ceph3=up:active}, 2 up:standby

     osdmap e389: 36 osds: 36 up, 36 in

      pgmap v19370: 8256 pgs, 3 pools, 51217 MB data, 14035 objects

            95526 MB used, 196 TB / 196 TB avail

                8256 active+clean

 

Our Ceph.conf is relatively simple at the moment:

 

cat /etc/ceph/ceph.conf

[global]

fsid = 4ed5ecdd-0c5b-4422-9d99-c9e42c6bd4cd

mon_initial_members = ceph1, ceph2, ceph3

mon_host = 10.15.0.1,10.15.0.2,10.15.0.3

mon_pg_warn_max_per_osd = 1000

auth_cluster_required = cephx

auth_service_required = cephx

auth_client_required = cephx

filestore_xattr_use_omap = true

osd_pool_default_size = 2

 

When I pulled the plug on the master MDS last time (ceph1), it stopped all IO until I plugged it back in. I was under the assumption that the MDS would fail over the other 2 MDS’s and IO would continue?

 

Is there something I need to do to allow the MDS’s to failover from each other without too much interruption? Or is this because the clients ceph version?

 

Cheers,

 

Simon Hallam

Linux Support & Development Officer

 

Please visit our new website at www.pml.ac.uk and follow us on Twitter  @PlymouthMarine

Winner of the Environment & Conservation category, the Charity Awards 2014.

Plymouth Marine Laboratory (PML) is a company limited by guarantee registered in England & Wales, company number 4178503. Registered Charity No. 1091222. Registered Office: Prospect Place, The Hoe, Plymouth  PL1 3DH, UK. 

This message is private and confidential. If you have received this message in error, please notify the sender and remove it from your system. You are reminded that e-mail communications are not secure and may contain viruses; PML accepts no liability for any loss or damage which may be caused by viruses.
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux