Re: mdadm udev rule does not start mdmonitor systemd unit.

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Nov 1, 2022 at 8:27 PM Marc Rechté <marc4@xxxxxxxxx> wrote:
>
> Hello,
>
> I have a udev rule and a md127 device with the properties as following.
>
> The mdmonitor service is not started (no trace in systemd journal).
> However I can manually start the service.
>
> I just noticed that SYSTEMD_READY porperty is 0 which could explain this
> behaviour (according to man systemd.device) ?

Hi Marc

For raid device, SYSTEMD_READY will be 1 when the change event happens.
And for lvm volume, SYSTEMD_READY will be 1 when the add event happens.
So you need to notice about his in your udev rule.

>
> I don't know how to further debug.

You can add systemd.log_level=debug udev.log-priority=debug to your boot conf
file. For example,
/boot/loader/entries/xxx-4.18.0-416.el8.x86_64.conf. My environment
is rhel. Maybe it's different on your system.

Then you can add some printf logs into your udev rule. I did in this
way, something
like this:

ENV{SYSTEMD_READY}=="0", GOTO="test_end"
SUBSYSTEM=="block", ACTION=="add", RUN{program}+="/usr/bin/echo
mdadm-test-add-SYSTEMD_READY"
SUBSYSTEM=="block", ACTION=="change", RUN{program}+="/usr/bin/echo
mdadm-test-change-SYSTEMD_READY"

You can check the logs by journalctl command. So you can know which
rule runs in your udev rule.

Regards
Xiao
>
> Thanks
>
> # udevadm info --query=property --name=/dev/md127
>
> DEVPATH=/devices/virtual/block/md127
> DEVNAME=/dev/md127
> DEVTYPE=disk
> DISKSEQ=6
> MAJOR=9
> MINOR=127
> SUBSYSTEM=block
> USEC_INITIALIZED=5129215
> ID_IGNORE_DISKSEQ=1
> MD_LEVEL=raid1
> MD_DEVICES=2
> MD_METADATA=1.2
> MD_UUID=800ee577:652e6fdc:79f6768e:dea2f7ea
> MD_DEVNAME=SysRAID1Array1
> MD_NAME=linux2:SysRAID1Array1
> ID_FS_UUID=x94VGG-7hfP-rn1c-MR53-q6to-QPZR-73eAdq
> ID_FS_UUID_ENC=x94VGG-7hfP-rn1c-MR53-q6to-QPZR-73eAdq
> ID_FS_VERSION=LVM2 001
> ID_FS_TYPE=LVM2_member
> ID_FS_USAGE=raid
> SYSTEMD_WANTS=mdmonitor.service
> SYSTEMD_READY=0
> UDISKS_MD_LEVEL=raid1
> UDISKS_MD_DEVICES=2
> UDISKS_MD_METADATA=1.2
> UDISKS_MD_UUID=800ee577:652e6fdc:79f6768e:dea2f7ea
> UDISKS_MD_DEVNAME=SysRAID1Array1
> UDISKS_MD_NAME=linux2:SysRAID1Array1
> UDISKS_MD_DEVICE_dev_nvme0n1p2_ROLE=0
> UDISKS_MD_DEVICE_dev_nvme0n1p2_DEV=/dev/nvme0n1p2
> UDISKS_MD_DEVICE_dev_nvme1n1p2_ROLE=1
> UDISKS_MD_DEVICE_dev_nvme1n1p2_DEV=/dev/nvme1n1p2
> DEVLINKS=/dev/md/SysRAID1Array1
> /dev/disk/by-id/md-name-linux2:SysRAID1Array1
> /dev/disk/by-id/lvm-pv-uuid-x94VGG-7hfP-rn1c-MR53-q6to-QPZR-73eAdq
> /dev/disk/by-id/md-uuid-800ee577:652e6fdc:79f6768e:dea2f7ea
> TAGS=:systemd:
> CURRENT_TAGS=:systemd:
>
> # cat /usr/lib/udev/rules.d/63-md-raid-arrays.rules
> # do not edit this file, it will be overwritten on update
>
> SUBSYSTEM!="block", GOTO="md_end"
>
> # handle md arrays
> ACTION!="add|change", GOTO="md_end"
> KERNEL!="md*", GOTO="md_end"
>
> # partitions have no md/{array_state,metadata_version}, but should not
> # for that reason be ignored.
> ENV{DEVTYPE}=="partition", GOTO="md_ignore_state"
>
> # container devices have a metadata version of e.g. 'external:ddf' and
> # never leave state 'inactive'
> ATTR{md/metadata_version}=="external:[A-Za-z]*",
> ATTR{md/array_state}=="inactive", GOTO="md_ignore_state"
> TEST!="md/array_state", ENV{SYSTEMD_READY}="0", GOTO="md_end"
> ATTR{md/array_state}=="clear*|inactive", ENV{SYSTEMD_READY}="0",
> GOTO="md_end"
> ATTR{md/sync_action}=="reshape", ENV{RESHAPE_ACTIVE}="yes"
> LABEL="md_ignore_state"
>
> IMPORT{program}="/usr/bin/mdadm --detail --no-devices --export $devnode"
> ENV{DEVTYPE}=="disk", ENV{MD_NAME}=="?*",
> SYMLINK+="disk/by-id/md-name-$env{MD_NAME}",
> OPTIONS+="string_escape=replace"
> ENV{DEVTYPE}=="disk", ENV{MD_UUID}=="?*",
> SYMLINK+="disk/by-id/md-uuid-$env{MD_UUID}"
> ENV{DEVTYPE}=="disk", ENV{MD_DEVNAME}=="?*", SYMLINK+="md/$env{MD_DEVNAME}"
> ENV{DEVTYPE}=="partition", ENV{MD_NAME}=="?*",
> SYMLINK+="disk/by-id/md-name-$env{MD_NAME}-part%n",
> OPTIONS+="string_escape=replace"
> ENV{DEVTYPE}=="partition", ENV{MD_UUID}=="?*",
> SYMLINK+="disk/by-id/md-uuid-$env{MD_UUID}-part%n"
> ENV{DEVTYPE}=="partition", ENV{MD_DEVNAME}=="*[^0-9]",
> SYMLINK+="md/$env{MD_DEVNAME}%n"
> ENV{DEVTYPE}=="partition", ENV{MD_DEVNAME}=="*[0-9]",
> SYMLINK+="md/$env{MD_DEVNAME}p%n"
>
> IMPORT{builtin}="blkid"
> OPTIONS+="link_priority=100"
> OPTIONS+="watch"
> ENV{ID_FS_USAGE}=="filesystem|other|crypto", ENV{ID_FS_UUID_ENC}=="?*",
> SYMLINK+="disk/by-uuid/$env{ID_FS_UUID_ENC}"
> ENV{ID_FS_USAGE}=="filesystem|other", ENV{ID_PART_ENTRY_UUID}=="?*",
> SYMLINK+="disk/by-partuuid/$env{ID_PART_ENTRY_UUID}"
> ENV{ID_FS_USAGE}=="filesystem|other", ENV{ID_FS_LABEL_ENC}=="?*",
> SYMLINK+="disk/by-label/$env{ID_FS_LABEL_ENC}"
>
> ENV{MD_LEVEL}=="raid[1-9]*", ENV{SYSTEMD_WANTS}+="mdmonitor.service"
>
> # Tell systemd to run mdmon for our container, if we need it.
> ENV{MD_LEVEL}=="raid[1-9]*", ENV{MD_CONTAINER}=="?*",
> PROGRAM="/usr/bin/readlink $env{MD_CONTAINER}", ENV{MD_MON_THIS}="%c"
> ENV{MD_MON_THIS}=="?*", PROGRAM="/usr/bin/basename $env{MD_MON_THIS}",
> ENV{SYSTEMD_WANTS}+="mdmon@%c.service"
> ENV{RESHAPE_ACTIVE}=="yes", PROGRAM="/usr/bin/basename
> $env{MD_MON_THIS}", ENV{SYSTEMD_WANTS}+="mdadm-grow-continue@%c.service"
>
> LABEL="md_end"
>
>





[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux