Re: Do you see a data loss if a SSD hosting several OSD journals crashes

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

Yes and no, for the actual data loss.  This depends on your crush map.

If you're using the original map (which came with the installation), then your smallest failure domain will be the host. If you have replica size and 3 hosts and 5 OSDs per host (15 OSDs total), then loosing the journal SSD in one host will only result in the data on that specific host being lost, and having to be re-created from the other two hosts. This is an anticipated failure.

If you changed the crush map to make the smallest failure domain to be the OSD, and ceph places all copies of a piece of data on the OSDs belonging to the SAME journal, then yes, you could end up with some pieces of data completely lost when that journal dies.

If I were in your shoes and I didn't want to set the the smallest failure domain to be the host, then I would create a new level 'ssd' and make that my failure domain. This way, if I had 10 OSDs and 2 SSD Journals per host, my crush map would look like this: 5 hosts -> 2 Journals/host -> 5 OSDs/Journal. This way, if I lost a journal, I would be losing only one copy of my data. Ceph would not place more than one copy of data per Journal (even though there are 5 OSDs behind that Journal).

I know this is a bit advanced, but I hope this clarifies things for you.

Dyweni




On 2016-05-19 20:59, Christian Balzer wrote:
Hello,

first of all, wall of text. Don't do that.
Use returns and paragraphs liberally to make reading easy.
I'm betting at least half of the people who could have answered you
question took a look at this blob of text and ignored it.

Secondly, search engines are your friend.
The first hit when googling for "ceph ssd journal failure" is this gem:
http://ceph.com/planet/ceph-recover-osds-after-ssd-journal-failure/

Loosing a journal SSD will at most cost you the data on all associated
OSDs and thus the recovery/backfill traffic, if you don't feel like doing
what the link above describes.

Ceph will not acknowledge a client write before all journals (replica
size, 3 by default) have received the data, so loosing one journal SSD
will NEVER result in an actual data loss.

Christian

On Fri, 20 May 2016 01:38:08 +0000 EP Komarla wrote:

  *   We are trying to assess if we are going to see a data loss if an
SSD that is hosting journals for few OSDs crashes. In our configuration,
each SSD is partitioned into 5 chunks and each chunk is mapped as a
journal drive for one OSD. What I understand from the Ceph
documentation: "Consistency: Ceph OSD Daemons require a filesystem
interface that guarantees atomic compound operations. Ceph OSD Daemons
write a description of the operation to the journal and apply the
operation to the filesystem. This enables atomic updates to an object
(for example, placement group metadata). Every few seconds-between
filestore max sync interval and filestore min sync interval-the Ceph OSD
Daemon stops writes and synchronizes the journal with the filesystem,
allowing Ceph OSD Daemons to trim operations from the journal and reuse
the space. On failure, Ceph OSD Daemons replay the journal starting
after the last synchronization operation." So, my question is what
happens if an SSD fails - am I going to lose all the data that has not
been written/synchronized to OSD?  In my case, am I going to lose data
for all the 5 OSDs which can be bad? This is of concern to us. What are
the options to prevent any data loss at all?  Is it better to have the
journals on the same hard drive, i.e., to have one journal per OSD and
host it on the same hard drive?  Of course, performance will not be as
good as having an SSD for OSD journal. In this case, I am thinking I
will not lose data as there are secondary OSDs where data is replicated (we are using triple replication). Any thoughts? What other solutions people have adopted for data reliability and consistency to address the
case I am mentioning?



Legal Disclaimer:
The information contained in this message may be privileged and
confidential. It is intended to be read only by the individual or entity
to whom it is addressed or by their designee. If the reader of this
message is not the intended recipient, you are on notice that any
distribution of this message, in any form, is strictly prohibited. If
you have received this message in error, please immediately notify the
sender and delete or destroy any copy of this message!


--
Christian Balzer        Network/Systems Engineer
chibi@xxxxxxx   	Global OnLine Japan/Rakuten Communications
http://www.gol.com/
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux