Re: Wish list : automatic rebuild with hot swap osd ?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thu, Oct 19, 2017 at 7:08 AM, Willem Jan Withagen <wjw@xxxxxxxxxxx> wrote:
> On 19-10-2017 14:14, Alfredo Deza wrote:
>>
>> If these properties also mean device information (vendor, size,
>> solid/rotational, etc...) it could help
>> to better map/detect an OSD replacement since clusters tend to have a
>> certain level of
>> homogeneous hardware: if $brand, and $size, and $rotational etc...
>>
>>>
>>> - A daemon (e.g., ceph-osd-autoreplace) that runs on each machine or a
>>> tool that is triggered by udev.  It would check for new, empty devices
>>> appearing in the locations (as defined by the by-path string) previously
>>> occupied by OSDs that are down.  If that happens, it can use 'ceph osd
>>> safe-to-destroy' to verify whether it is safe to automatically rebuild
>>> that OSD.  (If not, it might want to raise a health alert, since it's
>>> possible the drive that was physically pulled should be preserved until
>>> the cluster is sure it doesn't need it.)
>>
>>
>> systemd has some support for devices, so we might not even need a
>> daemon, but more a unit that can
>> depend on events already handled by systemd (would save us from udev).
>
>
> FreeBSD does not have systemd. 8-)
>
> I'm inclined to say luckily, but then that may be my personal bias.
> I don't like "automagic" tools like Udev or systemd tinkering with my disks.
>
> As Alan says, in ZFS one can designate hot-standby. But even there I prefer
> to be alerted and then manually intervene.

Actually, I was talking about autoreplace by physical path.  Hot
spares are something else.  The physical path of a drive is distinct
from its device path.  The physical path is determined by information
from a SES[1] expander, which can actually tell you which physical
slots contain which logical drives.

>
> A hot-swap daemon that gets instructed to only use explicitly and fully
> enumerated disk might be something to trust. So something matching
> disk-serial number would be oke.

Matching disk serial number isn't always safe in a VM.  VMs can
generate duplicate serial numbers.  Better to match against a GPT
label or something that identifies a drive as belonging to Ceph.
That, unfortunately, requires some intervention from the
administrator.  The nice thing about a user space daemon is that its
behavior can easily be controlled by the sysadmin.  So for example, a
sysadmin could opt into a rule that says "Ceph can take over all SCSI
disks" or "Ceph can take over all disks without an existing partition
table or known filesystem".

-Alan

[1] https://en.wikipedia.org/wiki/SCSI_Enclosure_Services
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux