Re: Ceph storage project for virtualization

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Eneko! 

Sorry for the delay answering. Thank you so much for your time really
mate :) . I answer you in green bold for instance between your lines for
better understanding what I'm talking about.

moving forward below! 

El 2024-03-05 12:26, Eneko Lacunza escribió:

> Hi Egoitz,
> 
> I don't think it is a good idea, but can't comment about if that's possible because I don't know well enough Ceph's inner workings, maybe others can comment.
> 
> This is what worries me:
> "
> 
> ach NFS redundant service of each datacenter will be composed by two
> NFS gateways accessing to the OSDs of the placement group located in the
> own datacenter. I planned achieving this with OSD weights and getting
> with that the fact that the crush algorithm to build the map so that
> each datacenter accesses end up having as master, the OSD of the own
> datacenter in the placement group. Obviously, slave OSD replicas will
> exist in the other three datacenters or even I don't discard the fact of
> using erasure coding in some manner.
> 
> " 
> 
> First, I don't think you got OSD weights right. 
> 
> YOU MEAN I DON'T HAVE THE CONCEPT CLEAR OR... THAT WEIGHTS DOESN'T WORK THAT WAY OR... PLEASE TELL ME FOR UNDERSTANDING WHAT YOU MEAN :) . IN CASE I DON'T HAVE THE CONCEPT CLEAR :) I'LL GO BACK TO READ ABOUT IT :) 
> 
> Also, any write will be synchronous to the replicas so that's why I asked about latencies first. You may be able to read from DC-local "master" pgs (I recall someone doing this with host-local pgs...) 
> 
> YOU MEAN HERE THAT AS I/O IS SYNCHRONOUS THE LATENCY IS EXTREMELY IMPORTANT WETHER YOU JUST ACCESS TO THE SAME DATACENTER OR NOT FROM THAT DATACENTER HIPERVISORS....
> 
> In the best case you'll have your data in a corner-case configuration, which may trigger strange bugs and/or behaviour not seen elsewhere. 
> 
> I wouldn't like to be in such a position, but I don't know how valuable your data is... 
> 
> I SEE... YES THE INFO IS ESSENTIAL REALLY.... 
> 
> PERHAPS THEN I WOULD HAVE ALL THE OSD SERVERS IN THE SAME DATACENTER FOR AVOIDING THAT DELAYS.... AND THAT EXTRANGE ISSUES THAT COULD HAPPEN WITH MY ORIGINAL IDEA...
> 
> I think it would be best to determine inter-DC network latency first; if you can choose DCs, then choose wisely with low enough latency ;) Then see if a regular Ceph storage configuration will give you good enough performance. 
> 
> UNDERSTOOD YES!! 
> 
> Another option would be to run DC-local ceph storages and to mirror to other DC. 
> 
> THIS IS VALID TOO... ALTHOUGH THAT WOULD BE SYNCHRONOUS?. I MEAN THE MIRROR?.
> 
> Cheers 
> 
> THANKS A LOT MATE!! REALLY!!!
> 
> El 5/3/24 a las 11:50, egoitz@xxxxxxxxxxxxx escribió: Hi Eneko!
> 
> I don't really have that data but I was planning to have as master OSD
> only the ones in the same datacenter as the hypervisor using the
> storage. The other datacenters would be just replicas. I assume you ask
> it because replication is totally synchronous.
> 
> Well for doing step by step. Imagine for the moment, the point of
> failure is a rack and all the replicas will be in the same datacenter in
> different racks and rows. In this case the latency should be acceptable
> and low.
> 
> My question was more related to the redundant nfs and if you have some
> experience with similar setups. I was trying to know if first is
> feasible what I'm planning to do.
> 
> Thank you so much :)
> 
> Cheers!
> 
> El 2024-03-05 11:43, Eneko Lacunza escribió:
> 
> Hi Egoitz,
> 
> What network latency between datacenters?
> 
> Cheers
> 
> El 5/3/24 a las 11:31, egoitz@xxxxxxxxxxxxx escribió:
> 
> Hi!
> 
> I have been reading some ebooks of Ceph and some doc and learning about
> it. The goal of all it, is the fact of creating a rock solid storage por
> virtual machines. After all the learning I have not been able to answer
> by myself to this question so I was wondering if perhaps you could
> clarify my doubt.
> 
> Let's imagine three datacenters, each one with for instance, 4
> virtualization hosts. As I was planning to build a solution for diferent
> hypervisors I have been thinking in the following env.
> 
> - I planed to have my Ceph storage (with different pools inside) with
> OSDs in three different datacenters (as failure point).
> 
> - Each datacenter's hosts, will be accessing to a NFS redundant service
> in the own datacenter.
> 
> - Each NFS redundant service of each datacenter will be composed by two
> NFS gateways accessing to the OSDs of the placement group located in the
> own datacenter. I planned achieving this with OSD weights and getting
> with that the fact that the crush algorithm to build the map so that
> each datacenter accesses end up having as master, the OSD of the own
> datacenter in the placement group. Obviously, slave OSD replicas will
> exist in the other three datacenters or even I don't discard the fact of
> using erasure coding in some manner.
> 
> - The NFS gateways could be a NFS redundant gateway service from Ceph (I
> have seen now they have developed something for this purpose
> https://docs.ceph.com/en/quincy/mgr/nfs/) or perhaps two different
> Debian machines, accessing to Ceph with rados and sharing to the
> hypervisors that information over NFS. In case of Debian machines I have
> heard good results using pacemaker/corosync for providing HA to that NFS
> (between 0,5 and 3 seconds for fail over and service up again).
> 
> What do you think about this plan?. Do you see it feasible?. We will
> work too with KVM and there we could access to Ceph directly but I would
> needed to provide too storage por Xen and Vmware.
> 
> Thank you so much in advance,
> 
> Cheers!
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx Eneko Lacunza
> Zuzendari teknikoa | Director técnico
> Binovo IT Human Project
> 
> Tel. +34 943 569 206 | https://www.binovo.es
> Astigarragako Bidea, 2 - 2º izda. Oficina 10-11, 20180 Oiartzun
> 
> https://www.youtube.com/user/CANALBINOVO
> https://www.linkedin.com/company/37269706/
 _______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx 
Eneko Lacunza
Zuzendari teknikoa | Director técnico
Binovo IT Human Project

Tel. +34 943 569 206 | https://www.binovo.es
Astigarragako Bidea, 2 - 2º izda. Oficina 10-11, 20180 Oiartzun

https://www.youtube.com/user/CANALBINOVO
https://www.linkedin.com/company/37269706/
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux