Re: Ceph mount rbd

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Gonzalo,

 

You are right, i told so much about my enviroment actual and maybe i didn't know explain my problem the better form, with ceph in the moment, mutiple hosts clients can mount and write datas in my system and this is one problem, because i could have filesystem corruption.

 

Example, today, if runing the comand in two machines in the same time, it will work.

 

mount /dev/rbd0 /mnt/veeamrepo

cd /mnt/veeamrepo ; touch testfile.txt

 

I need ensure, only one machine will can execute this.

 

Thanks a lot,

Marcelo


Em 14/07/2017, Gonzalo Aguilar Delgado <gaguilar@xxxxxxxxxxxxxxxxxx>
escreveu:
> Hi,
>
> Why you would like to maintain copies by yourself. You replicate on ceph
> and then on different files inside ceph? Let ceph take care of counting.
> Create a pool with 3 or more copies and let ceph take care of what's
> stored and where.
>
> Best regards,
>
>
> El 13/07/17 a las 17:06, lista@xxxxxxxxxxxxxxxxx escribió:
> >
> > I will explain More about my system actual, in the moment i have 2
> > machines using drbd in mode master/slave and i running the aplication
> > in machine master, but existing 2 questions importants in my
> > enviroment with drbd actualy :
> >
> > 1 - If machine one is master and mounting partitions, the slave don't
> > can mount the system, Unless it happens one problem in machine master,
> > this is one mode, to prevent write in filesystem incorrect
> >
> > 2 - When i write data in machine master in drbd, the drbd write datas
> > in slave machine Automatically, with this, if one problem happens in
> > node master, the machine slave have coppy the data.
> >
> > In the moment, in my enviroment testing with ceph, using the version
> > 4.10 of kernel and i mount the system in two machines in the same
> > time, in production enviroment, i could serious problem with this
> > comportament.
> >
> > How can i use the ceph and Ensure that I could get these 2 behaviors
> > kept in a new environment with Ceph?
> >
> > Thanks a lot,
> >
> > Marcelo
> >
> >
> > Em 28/06/2017, Jason Dillaman <jdillama@xxxxxxxxxx> escreveu:
> > > ... additionally, the forthcoming 4.12 kernel release will support
> > > non-cooperative exclusive locking. By default, since 4.9, when the
> > > exclusive-lock feature is enabled, only a single client can write to
> > the
> > > block device at a time -- but they will cooperatively pass the lock
> > back
> > > and forth upon write request. With the new "rbd map" option, you can
> > map a
> > > image on exactly one host and prevent other hosts from mapping the
> > image.
> > > If that host should die, the exclusive-lock will automatically become
> > > available to other hosts for mapping.
> > >
> > > Of course, I always have to ask the use-case behind mapping the same
> > image
> > > on multiple hosts. Perhaps CephFS would be a better fit if you are
> > trying
> > > to serve out a filesystem?
> > >
> > > On Wed, Jun 28, 2017 at 6:25 PM, Maged Mokhtar
> > <mmokhtar@xxxxxxxxxxx> wrote:
> > >
> > > > On 2017-06-28 22:55, lista@xxxxxxxxxxxxxxxxx wrote:
> > > >
> > > > Hi People,
> > > >
> > > > I am testing the new enviroment, with ceph + rbd with ubuntu
> > 16.04, and i
> > > > have one question.
> > > >
> > > > I have my cluster ceph and mount the using the comands to ceph in
> > my linux
> > > > enviroment :
> > > >
> > > > rbd create veeamrepo --size 20480
> > > > rbd --image veeamrepo info
> > > > modprobe rbd
> > > > rbd map veeamrepo
> > > > rbd feature disable veeamrepo exclusive-lock object-map fast-diff
> > > > deep-flatten
> > > > mkdir /mnt/veeamrepo
> > > > mount /dev/rbd0 /mnt/veeamrepo
> > > >
> > > > The comands work fine, but i have one problem, in the moment, i
> > can mount
> > > > the /mnt/veeamrepo in the same time in 2 machines, and this is a
> > bad option
> > > > for me in the moment, because this could generate one filesystem
> > corrupt.
> > > >
> > > > I need only one machine to be allowed to mount and write at a time.
> > > >
> > > > Example if machine1 mount the /mnt/veeamrepo and machine2 try
> > mount, one
> > > > error would be displayed, show message the machine can not mount,
> > because
> > > > the system already mounted in machine1.
> > > >
> > > > Someone, could help-me with this or give some tips, for solution my
> > > > problem. ?
> > > >
> > > > Thanks a lot
> > > >
> > > > _______________________________________________
> > > > ceph-users mailing list
> > > > ceph-users@xxxxxxxxxxxxxx
> > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> > > >
> > > >
> > > >
> > > > You can use Pacemaker to map the rbd and mount the filesystem on 1
> > server
> > > > and in case of failure switch to another server.
> > > >
> > > >
> > > > _______________________________________________
> > > > ceph-users mailing list
> > > > ceph-users@xxxxxxxxxxxxxx
> > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> > > >
> > > >
> > >
> > >
> > > --
> > > Jason
> >
> >
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@xxxxxxxxxxxxxx
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
> ----
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux