Hi Angapov,
I have seen below message in ceph official sites. How it's consider to use in production.
"ImportantCephFS currently lacks a robust ‘fsck’ check and repair function. Please use caution when storing important data as the disaster recovery tools are still under development. For more information about using CephFS today, see CephFS for early adopters"
Regards
Prabu
I have seen below message in ceph official sites. How it's consider to use in production.
"ImportantCephFS currently lacks a robust ‘fsck’ check and repair function. Please use caution when storing important data as the disaster recovery tools are still under development. For more information about using CephFS today, see CephFS for early adopters"
Regards
Prabu
Regards
G.J
---- On Thu, 21 May 2015 19:57:09 +0530 angapov@xxxxxxxxx wrote ----Hi, Prabu!This behavior is expected because you are using non-clustered filesystem (ext4 or xfs or whatever), which is not expected to be mounted to multiple hosts at the same time.What's more - you can lose data when doing like this. That's the nature of local filesystems.So if you need to access filesystem simultaneously at two or more hosts - consider using clustered filesystem like CephFS, OCFS etc or using network mounts like NFS or Samba.Regards, Vasily.On Thu, May 21, 2015 at 5:10 PM, gjprabu <gjprabu@xxxxxxxxxxxx> wrote:Hi All,
We are using rbd and map the same rbd image to the rbd device on two different client but i can't see the data until i umount and mount -a partition. Kindly share the solution for this issue.
Example
create rbd image named foo
map foo to /dev/rbd0 on server A, mount /dev/rbd0 to /mnt
map foo to /dev/rbd0 on server B, mount /dev/rbd0 to /mnt
Regards
Prabu
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com