Re: Basic Ceph questions

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Thanks:)

If someone can help reg the question below, that would be great!

"
>
> For VMs, I am trying to visualize how the RBD device would be exposed.
> Where does the driver live exactly? If its exposed via libvirt and
> QEMU, does the kernel driver run in the host OS, and communicate with
> a backend Ceph cluster? If yes, does libRBD provide a target (SCSI?)
> interface which the kernel driver connects to? Trying to visualize
> what the stack looks like, and the flow of IOs for block devices.
"

MW

On Fri, Oct 10, 2014 at 10:18 AM, Craig Lewis <clewis@xxxxxxxxxxxxxxxxxx> wrote:
>>
>> Just curious, what kind of applications use RBD? It cant be
>> applications which need high speed SAN storage performance
>> characteristics?
>
>
> Most people seem to be using it as storage for OpenStack.
>
> I've heard about people using RDB + Heartbeats to make an HA NFS, while they
> wait for CephFS to be production ready.
>
> People that are re-exporting images via iSCSI and Fiber Channel are probably
> doing something different.  If I had to hazard a guess, I'd guess that
> they're doing some sort of HA Clustered service, like a database.  That's
> the traditional use for shared storage.
>
>
>
>>
>>
>> For VMs, I am trying to visualize how the RBD device would be exposed.
>> Where does the driver live exactly? If its exposed via libvirt and
>> QEMU, does the kernel driver run in the host OS, and communicate with
>> a backend Ceph cluster? If yes, does libRBD provide a target (SCSI?)
>> interface which the kernel driver connects to? Trying to visualize
>> what the stack looks like, and the flow of IOs for block devices.
>
>
> I'll have to leave that for others to answer.
>
>
>> >>
>> >> b. If it is strongly consistent, is that the case across sites also?
>> >> How can it be performant across geo sites if that is the case? If its
>> >> choosing consistency over partitioning and availability...For object,
>> >> I read somewhere that it is now eventually consistent(local CP,
>> >> remotely AP) via DR. Gets a bit confusing with all the literature out
>> >> there. If it is DR, isnt that slightly different from the Swift case?
>> >
>> >
>> > If you're referring to RadosGW Federation, no.  That replication is
>> > async.
>> > The replication has several delays built in, so the fastest you could to
>> > see
>> > your data show up in the secondary is about a minute.  Longer if the
>> > file
>> > takes a while to transfer, or you have a lot of activity to replicate.
>> >
>> > Each site is still CP.  There is just delay getting data from the
>> > primary to
>> > the secondary.
>> In that case, it is like Swift, only differently done. The async makes
>> it eventually consistent across sites, no?
>
>
> I'm not sure regarding Swift.  Also outside my experience.
>
> But yes, the async replication is eventually consistent, with no guarantee.
> Problems during replication can cause the clusters to get out of sync.  The
> replication agent will retry failures, but it doesn't store the information
> anywhere.  If you restart the replication agent when it had known failures,
> those failures won't be retried.  Every one of the errors is logged, so I
> was able to manually download & re-upload the file to the primary cluster,
> which triggered re-replication.
>
> So far, all of the inconsistencies have shown up by comparing bucket
> listing.  I'm in the process of manually verifying checksums (my application
> stores a SHA256 for every object uploaded).  So far, I haven't had any
> failures in files that were marked as successfully replicated.
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux