Do you see a data loss if a SSD hosting several OSD journals crashes

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



  • We are trying to assess if we are going to see a data loss if an SSD that is hosting journals for few OSDs crashes. In our configuration, each SSD is partitioned into 5 chunks and each chunk is mapped as a journal drive for one OSD. What I understand from the Ceph documentation: “Consistency: Ceph OSD Daemons require a filesystem interface that guarantees atomic compound operations. Ceph OSD Daemons write a description of the operation to the journal and apply the operation to the filesystem. This enables atomic updates to an object (for example, placement group metadata). Every few seconds–between filestore max sync interval and filestore min sync interval–the Ceph OSD Daemon stops writes and synchronizes the journal with the filesystem, allowing Ceph OSD Daemons to trim operations from the journal and reuse the space. On failure, Ceph OSD Daemons replay the journal starting after the last synchronization operation.” So, my question is what happens if an SSD fails – am I going to lose all the data that has not been written/synchronized to OSD?  In my case, am I going to lose data for all the 5 OSDs which can be bad?  This is of concern to us. What are the options to prevent any data loss at all?  Is it better to have the journals on the same hard drive, i.e., to have one journal per OSD and host it on the same hard drive?  Of course, performance will not be as good as having an SSD for OSD journal. In this case, I am thinking I will not lose data as there are secondary OSDs where data is replicated (we are using triple replication).  Any thoughts?  What other solutions people have adopted for data reliability and consistency to address the case I am mentioning?

 

 


Legal Disclaimer:
The information contained in this message may be privileged and confidential. It is intended to be read only by the individual or entity to whom it is addressed or by their designee. If the reader of this message is not the intended recipient, you are on notice that any distribution of this message, in any form, is strictly prohibited. If you have received this message in error, please immediately notify the sender and delete or destroy any copy of this message!
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux