Hello,
I have a cluster CEPH (10.2.5-1trusty) I use the various possibilities:
-Block
- Object
- CephFS
root@ih-par1-cld1-ceph-01:~# cat /etc/ceph/ceph.conf [....] mon_host = 10.4.0.1, 10.4.0.3, 10.4.0.5 [....] public_network = 10.4.0.0/24 cluster_network = 192.168.33.0/24 [....]
I have dedicated servers for the storage Block, Object and other servers for CephFS (Full SSD):
root@ih-par1-cld1-ceph-01:~# ceph osd tree ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY -6 2.79593 root ssdforcephfs -7 0.46599 host ih-prd-cephfs-02 32 0.23299 osd.32 up 1.00000 1.00000 33 0.23299 osd.33 up 1.00000 1.00000 -8 0.46599 host ih-prd-cephfs-03 34 0.23299 osd.34 up 1.00000 1.00000 35 0.23299 osd.35 up 1.00000 1.00000 -9 0.46599 host ih-prd-cephfs-05 36 0.23299 osd.36 up 1.00000 1.00000 37 0.23299 osd.37 up 1.00000 1.00000 -10 0.46599 host ih-prd-cephfs-01 38 0.23299 osd.38 up 1.00000 1.00000 39 0.23299 osd.39 up 1.00000 1.00000 -11 0.46599 host ih-prd-cephfs-04 40 0.23299 osd.40 up 1.00000 1.00000 41 0.23299 osd.41 up 1.00000 1.00000 -12 0.46599 host ih-prd-cephfs-07 42 0.23299 osd.42 up 1.00000 1.00000 43 0.23299 osd.43 up 1.00000 1.00000 -1 116.47998 root default -2 43.67999 host ih-par1-cld1-ceph-01 0 3.64000 osd.0 up 1.00000 1.00000 2 3.64000 osd.2 up 1.00000 1.00000 6 3.64000 osd.6 up 1.00000 1.00000 8 3.64000 osd.8 up 1.00000 1.00000 15 3.64000 osd.15 up 1.00000 1.00000 16 3.64000 osd.16 up 1.00000 1.00000 19 3.64000 osd.19 up 1.00000 1.00000 22 3.64000 osd.22 up 1.00000 1.00000 24 3.64000 osd.24 up 1.00000 1.00000 26 3.64000 osd.26 up 1.00000 1.00000 28 3.64000 osd.28 up 1.00000 1.00000 30 3.64000 osd.30 up 1.00000 1.00000 -3 43.67999 host ih-par1-cld1-ceph-03 1 3.64000 osd.1 up 1.00000 1.00000 3 3.64000 osd.3 up 1.00000 1.00000 5 3.64000 osd.5 up 1.00000 1.00000 7 3.64000 osd.7 up 1.00000 1.00000 13 3.64000 osd.13 up 1.00000 1.00000 4 3.64000 osd.4 up 1.00000 1.00000 20 3.64000 osd.20 up 1.00000 1.00000 23 3.64000 osd.23 up 1.00000 1.00000 25 3.64000 osd.25 up 1.00000 1.00000 27 3.64000 osd.27 up 1.00000 1.00000 29 3.64000 osd.29 up 1.00000 1.00000 31 3.64000 osd.31 up 1.00000 1.00000 -5 29.12000 host ih-par1-cld1-ceph-05 9 3.64000 osd.9 up 1.00000 1.00000 10 3.64000 osd.10 up 1.00000 1.00000 11 3.64000 osd.11 up 1.00000 1.00000 12 3.64000 osd.12 up 1.00000 1.00000 14 3.64000 osd.14 up 1.00000 1.00000 17 3.64000 osd.17 up 1.00000 1.00000 18 3.64000 osd.18 up 1.00000 1.00000 21 3.64000 osd.21 up 1.00000 1.00000
I use OpenNebula for the use of RDB on the public network: 10.4.0.0/16.
I shall like separating the RDB, RGW network and CephFS... I have my customers CephFS who can accèder to all the network RBD, the hypervisors OpenNebula
Example:
Customer A (CephFS, path: /client1) = > Reaches at present all the network 10.4.0.0/16 Customer B (CephFS, path: /client2) = > Reaches at present all the network 10.4.0.0/16
How is it possible to separate networks: RBD, RGW and have multiple access networks for CephFS?
I hope to have been clear:/
Thank you
|
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com