On Fri, Jul 14, 2017 at 9:44 AM, <lista@xxxxxxxxxxxxxxxxx> wrote: > Gonzalo, > > > > You are right, i told so much about my enviroment actual and maybe i didn't > know explain my problem the better form, with ceph in the moment, mutiple > hosts clients can mount and write datas in my system and this is one > problem, because i could have filesystem corruption. > > > > Example, today, if runing the comand in two machines in the same time, it > will work. > > > > mount /dev/rbd0 /mnt/veeamrepo > > cd /mnt/veeamrepo ; touch testfile.txt > > > > I need ensure, only one machine will can execute this. > A user could do the same thing with any number of remote block devices (i.e. I could map an iSCSI target multiple times). As I said before, you can use the "exclusive" option available since kernel 4.12, roll your own solution using the advisory locks available from the rbd CLI, or just use CephFS if you want to be able to access a file system on multiple hosts. > > Thanks a lot, > > Marcelo > > > Em 14/07/2017, Gonzalo Aguilar Delgado <gaguilar@xxxxxxxxxxxxxxxxxx> > escreveu: > > >> Hi, >> >> Why you would like to maintain copies by yourself. You replicate on ceph >> and then on different files inside ceph? Let ceph take care of counting. >> Create a pool with 3 or more copies and let ceph take care of what's >> stored and where. >> >> Best regards, >> >> >> El 13/07/17 a las 17:06, lista@xxxxxxxxxxxxxxxxx escribió: >> > >> > I will explain More about my system actual, in the moment i have 2 >> > machines using drbd in mode master/slave and i running the aplication >> > in machine master, but existing 2 questions importants in my >> > enviroment with drbd actualy : >> > >> > 1 - If machine one is master and mounting partitions, the slave don't >> > can mount the system, Unless it happens one problem in machine master, >> > this is one mode, to prevent write in filesystem incorrect >> > >> > 2 - When i write data in machine master in drbd, the drbd write datas >> > in slave machine Automatically, with this, if one problem happens in >> > node master, the machine slave have coppy the data. >> > >> > In the moment, in my enviroment testing with ceph, using the version >> > 4.10 of kernel and i mount the system in two machines in the same >> > time, in production enviroment, i could serious problem with this >> > comportament. >> > >> > How can i use the ceph and Ensure that I could get these 2 behaviors >> > kept in a new environment with Ceph? >> > >> > Thanks a lot, >> > >> > Marcelo >> > >> > >> > Em 28/06/2017, Jason Dillaman <jdillama@xxxxxxxxxx> escreveu: >> > > ... additionally, the forthcoming 4.12 kernel release will support >> > > non-cooperative exclusive locking. By default, since 4.9, when the >> > > exclusive-lock feature is enabled, only a single client can write to >> > the >> > > block device at a time -- but they will cooperatively pass the lock >> > back >> > > and forth upon write request. With the new "rbd map" option, you can >> > map a >> > > image on exactly one host and prevent other hosts from mapping the >> > image. >> > > If that host should die, the exclusive-lock will automatically become >> > > available to other hosts for mapping. >> > > >> > > Of course, I always have to ask the use-case behind mapping the same >> > image >> > > on multiple hosts. Perhaps CephFS would be a better fit if you are >> > trying >> > > to serve out a filesystem? >> > > >> > > On Wed, Jun 28, 2017 at 6:25 PM, Maged Mokhtar >> > <mmokhtar@xxxxxxxxxxx> wrote: >> > > >> > > > On 2017-06-28 22:55, lista@xxxxxxxxxxxxxxxxx wrote: >> > > > >> > > > Hi People, >> > > > >> > > > I am testing the new enviroment, with ceph + rbd with ubuntu >> > 16.04, and i >> > > > have one question. >> > > > >> > > > I have my cluster ceph and mount the using the comands to ceph in >> > my linux >> > > > enviroment : >> > > > >> > > > rbd create veeamrepo --size 20480 >> > > > rbd --image veeamrepo info >> > > > modprobe rbd >> > > > rbd map veeamrepo >> > > > rbd feature disable veeamrepo exclusive-lock object-map fast-diff >> > > > deep-flatten >> > > > mkdir /mnt/veeamrepo >> > > > mount /dev/rbd0 /mnt/veeamrepo >> > > > >> > > > The comands work fine, but i have one problem, in the moment, i >> > can mount >> > > > the /mnt/veeamrepo in the same time in 2 machines, and this is a >> > bad option >> > > > for me in the moment, because this could generate one filesystem >> > corrupt. >> > > > >> > > > I need only one machine to be allowed to mount and write at a time. >> > > > >> > > > Example if machine1 mount the /mnt/veeamrepo and machine2 try >> > mount, one >> > > > error would be displayed, show message the machine can not mount, >> > because >> > > > the system already mounted in machine1. >> > > > >> > > > Someone, could help-me with this or give some tips, for solution my >> > > > problem. ? >> > > > >> > > > Thanks a lot >> > > > >> > > > _______________________________________________ >> > > > ceph-users mailing list >> > > > ceph-users@xxxxxxxxxxxxxx >> > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > > > >> > > > >> > > > >> > > > You can use Pacemaker to map the rbd and mount the filesystem on 1 >> > server >> > > > and in case of failure switch to another server. >> > > > >> > > > >> > > > _______________________________________________ >> > > > ceph-users mailing list >> > > > ceph-users@xxxxxxxxxxxxxx >> > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > > > >> > > > >> > > >> > > >> > > -- >> > > Jason >> > >> > >> > >> > _______________________________________________ >> > ceph-users mailing list >> > ceph-users@xxxxxxxxxxxxxx >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> ---- >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@xxxxxxxxxxxxxx >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > -- Jason _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com