Re: mdadm udev rule does not start mdmonitor systemd unit.

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Mon, Nov 7, 2022 at 4:48 PM Marc Rechté <marc4@xxxxxxxxx> wrote:
>
> Le 07/11/2022 à 09:30, Xiao Ni a écrit :
> > On Sun, Nov 6, 2022 at 4:51 PM Marc Rechté <marc4@xxxxxxxxx> wrote:
> >> Le 03/11/2022 à 03:54, Xiao Ni a écrit :
> >>> On Tue, Nov 1, 2022 at 8:27 PM Marc Rechté <marc4@xxxxxxxxx> wrote:
> >>>> Hello,
> >>>>
> >>>> I have a udev rule and a md127 device with the properties as following.
> >>>>
> >>>> The mdmonitor service is not started (no trace in systemd journal).
> >>>> However I can manually start the service.
> >>>>
> >>>> I just noticed that SYSTEMD_READY porperty is 0 which could explain this
> >>>> behaviour (according to man systemd.device) ?
> >>> Hi Marc
> >>>
> >>> For raid device, SYSTEMD_READY will be 1 when the change event happens.
> >>> And for lvm volume, SYSTEMD_READY will be 1 when the add event happens.
> >>> So you need to notice about his in your udev rule.
> >>>
> >>>> I don't know how to further debug.
> >>> You can add systemd.log_level=debug udev.log-priority=debug to your boot conf
> >>> file. For example,
> >>> /boot/loader/entries/xxx-4.18.0-416.el8.x86_64.conf. My environment
> >>> is rhel. Maybe it's different on your system.
> >>>
> >>> Then you can add some printf logs into your udev rule. I did in this
> >>> way, something
> >>> like this:
> >>>
> >>> ENV{SYSTEMD_READY}=="0", GOTO="test_end"
> >>> SUBSYSTEM=="block", ACTION=="add", RUN{program}+="/usr/bin/echo
> >>> mdadm-test-add-SYSTEMD_READY"
> >>> SUBSYSTEM=="block", ACTION=="change", RUN{program}+="/usr/bin/echo
> >>> mdadm-test-change-SYSTEMD_READY"
> >>>
> >>> You can check the logs by journalctl command. So you can know which
> >>> rule runs in your udev rule.
> >>>
> >>> Regards
> >>> Xiao
> >>>> Thanks
> >>>>
> >>>> # udevadm info --query=property --name=/dev/md127
> >>>>
> >>>> DEVPATH=/devices/virtual/block/md127
> >>>> DEVNAME=/dev/md127
> >>>> DEVTYPE=disk
> >>>> DISKSEQ=6
> >>>> MAJOR=9
> >>>> MINOR=127
> >>>> SUBSYSTEM=block
> >>>> USEC_INITIALIZED=5129215
> >>>> ID_IGNORE_DISKSEQ=1
> >>>> MD_LEVEL=raid1
> >>>> MD_DEVICES=2
> >>>> MD_METADATA=1.2
> >>>> MD_UUID=800ee577:652e6fdc:79f6768e:dea2f7ea
> >>>> MD_DEVNAME=SysRAID1Array1
> >>>> MD_NAME=linux2:SysRAID1Array1
> >>>> ID_FS_UUID=x94VGG-7hfP-rn1c-MR53-q6to-QPZR-73eAdq
> >>>> ID_FS_UUID_ENC=x94VGG-7hfP-rn1c-MR53-q6to-QPZR-73eAdq
> >>>> ID_FS_VERSION=LVM2 001
> >>>> ID_FS_TYPE=LVM2_member
> >>>> ID_FS_USAGE=raid
> >>>> SYSTEMD_WANTS=mdmonitor.service
> >>>> SYSTEMD_READY=0
> >>>> UDISKS_MD_LEVEL=raid1
> >>>> UDISKS_MD_DEVICES=2
> >>>> UDISKS_MD_METADATA=1.2
> >>>> UDISKS_MD_UUID=800ee577:652e6fdc:79f6768e:dea2f7ea
> >>>> UDISKS_MD_DEVNAME=SysRAID1Array1
> >>>> UDISKS_MD_NAME=linux2:SysRAID1Array1
> >>>> UDISKS_MD_DEVICE_dev_nvme0n1p2_ROLE=0
> >>>> UDISKS_MD_DEVICE_dev_nvme0n1p2_DEV=/dev/nvme0n1p2
> >>>> UDISKS_MD_DEVICE_dev_nvme1n1p2_ROLE=1
> >>>> UDISKS_MD_DEVICE_dev_nvme1n1p2_DEV=/dev/nvme1n1p2
> >>>> DEVLINKS=/dev/md/SysRAID1Array1
> >>>> /dev/disk/by-id/md-name-linux2:SysRAID1Array1
> >>>> /dev/disk/by-id/lvm-pv-uuid-x94VGG-7hfP-rn1c-MR53-q6to-QPZR-73eAdq
> >>>> /dev/disk/by-id/md-uuid-800ee577:652e6fdc:79f6768e:dea2f7ea
> >>>> TAGS=:systemd:
> >>>> CURRENT_TAGS=:systemd:
> >>>>
> >>>> # cat /usr/lib/udev/rules.d/63-md-raid-arrays.rules
> >>>> # do not edit this file, it will be overwritten on update
> >>>>
> >>>> SUBSYSTEM!="block", GOTO="md_end"
> >>>>
> >>>> # handle md arrays
> >>>> ACTION!="add|change", GOTO="md_end"
> >>>> KERNEL!="md*", GOTO="md_end"
> >>>>
> >>>> # partitions have no md/{array_state,metadata_version}, but should not
> >>>> # for that reason be ignored.
> >>>> ENV{DEVTYPE}=="partition", GOTO="md_ignore_state"
> >>>>
> >>>> # container devices have a metadata version of e.g. 'external:ddf' and
> >>>> # never leave state 'inactive'
> >>>> ATTR{md/metadata_version}=="external:[A-Za-z]*",
> >>>> ATTR{md/array_state}=="inactive", GOTO="md_ignore_state"
> >>>> TEST!="md/array_state", ENV{SYSTEMD_READY}="0", GOTO="md_end"
> >>>> ATTR{md/array_state}=="clear*|inactive", ENV{SYSTEMD_READY}="0",
> >>>> GOTO="md_end"
> >>>> ATTR{md/sync_action}=="reshape", ENV{RESHAPE_ACTIVE}="yes"
> >>>> LABEL="md_ignore_state"
> >>>>
> >>>> IMPORT{program}="/usr/bin/mdadm --detail --no-devices --export $devnode"
> >>>> ENV{DEVTYPE}=="disk", ENV{MD_NAME}=="?*",
> >>>> SYMLINK+="disk/by-id/md-name-$env{MD_NAME}",
> >>>> OPTIONS+="string_escape=replace"
> >>>> ENV{DEVTYPE}=="disk", ENV{MD_UUID}=="?*",
> >>>> SYMLINK+="disk/by-id/md-uuid-$env{MD_UUID}"
> >>>> ENV{DEVTYPE}=="disk", ENV{MD_DEVNAME}=="?*", SYMLINK+="md/$env{MD_DEVNAME}"
> >>>> ENV{DEVTYPE}=="partition", ENV{MD_NAME}=="?*",
> >>>> SYMLINK+="disk/by-id/md-name-$env{MD_NAME}-part%n",
> >>>> OPTIONS+="string_escape=replace"
> >>>> ENV{DEVTYPE}=="partition", ENV{MD_UUID}=="?*",
> >>>> SYMLINK+="disk/by-id/md-uuid-$env{MD_UUID}-part%n"
> >>>> ENV{DEVTYPE}=="partition", ENV{MD_DEVNAME}=="*[^0-9]",
> >>>> SYMLINK+="md/$env{MD_DEVNAME}%n"
> >>>> ENV{DEVTYPE}=="partition", ENV{MD_DEVNAME}=="*[0-9]",
> >>>> SYMLINK+="md/$env{MD_DEVNAME}p%n"
> >>>>
> >>>> IMPORT{builtin}="blkid"
> >>>> OPTIONS+="link_priority=100"
> >>>> OPTIONS+="watch"
> >>>> ENV{ID_FS_USAGE}=="filesystem|other|crypto", ENV{ID_FS_UUID_ENC}=="?*",
> >>>> SYMLINK+="disk/by-uuid/$env{ID_FS_UUID_ENC}"
> >>>> ENV{ID_FS_USAGE}=="filesystem|other", ENV{ID_PART_ENTRY_UUID}=="?*",
> >>>> SYMLINK+="disk/by-partuuid/$env{ID_PART_ENTRY_UUID}"
> >>>> ENV{ID_FS_USAGE}=="filesystem|other", ENV{ID_FS_LABEL_ENC}=="?*",
> >>>> SYMLINK+="disk/by-label/$env{ID_FS_LABEL_ENC}"
> >>>>
> >>>> ENV{MD_LEVEL}=="raid[1-9]*", ENV{SYSTEMD_WANTS}+="mdmonitor.service"
> >>>>
> >>>> # Tell systemd to run mdmon for our container, if we need it.
> >>>> ENV{MD_LEVEL}=="raid[1-9]*", ENV{MD_CONTAINER}=="?*",
> >>>> PROGRAM="/usr/bin/readlink $env{MD_CONTAINER}", ENV{MD_MON_THIS}="%c"
> >>>> ENV{MD_MON_THIS}=="?*", PROGRAM="/usr/bin/basename $env{MD_MON_THIS}",
> >>>> ENV{SYSTEMD_WANTS}+="mdmon@%c.service"
> >>>> ENV{RESHAPE_ACTIVE}=="yes", PROGRAM="/usr/bin/basename
> >>>> $env{MD_MON_THIS}", ENV{SYSTEMD_WANTS}+="mdadm-grow-continue@%c.service"
> >>>>
> >>>> LABEL="md_end"
> >>>>
> >>>>
> >> Hello Xiao,
> >>
> >> Thanks for the tips.
> >>
> >> It appears that SYSTEMD_READY == 1 when entering the add/change event,
> >> but it seems it is reset to 0 while processing the rules.
> >>
> >> Following is modified rule with debug info. Relevant journal entries:
> >>
> >> md127: '/usr/bin/echo mdadm-test-add-SYSTEMD_READY'(out)
> >> 'mdadm-test-add-SYSTEMD_READY'
> > You see the log in your test. From the following udev rule, it only can handle
> > add/change events. And SYSTEMD_READY is 1, it doesn't go to the md_test
> > label. And the log appears. So it doesn't reset to 0, right?
> >
> > Regards
> > Xiao
> >
> >> ...
> >>
> >> md127: '/usr/bin/udevadm info --query=property --name=/dev/md127'(out)
> >> 'SYSTEMD_READY=0'
> >>
> >>
> >> $ cat 63-md-raid-arrays.rules
> >>
> >> # do not edit this file, it will be overwritten on update
> >>
> >> SUBSYSTEM!="block", GOTO="md_end"
> >>
> >> # handle md arrays
> >> ACTION!="add|change", GOTO="md_end"
> >> KERNEL!="md*", GOTO="md_end"
> >>
> >> ENV{SYSTEMD_READY}=="0", GOTO="md_test"
> >> RUN{program}+="/usr/bin/echo mdadm-test-add-SYSTEMD_READY"
> >> LABEL="md_test"
> >>
> >>
> >> # partitions have no md/{array_state,metadata_version}, but should not
> >> # for that reason be ignored.
> >> ENV{DEVTYPE}=="partition", GOTO="md_ignore_state"
> >>
> >> # container devices have a metadata version of e.g. 'external:ddf' and
> >> # never leave state 'inactive'
> >> ATTR{md/metadata_version}=="external:[A-Za-z]*",
> >> ATTR{md/array_state}=="inactive", GOTO="md_ignore_state"
> >> TEST!="md/array_state", ENV{SYSTEMD_READY}="0", GOTO="md_end"
> >> ATTR{md/array_state}=="clear*|inactive", ENV{SYSTEMD_READY}="0",
> >> GOTO="md_end"
> >> ATTR{md/sync_action}=="reshape", ENV{RESHAPE_ACTIVE}="yes"
> >> LABEL="md_ignore_state"
> >>
> >> IMPORT{program}="/usr/bin/mdadm --detail --no-devices --export $devnode"
> >> ENV{DEVTYPE}=="disk", ENV{MD_NAME}=="?*",
> >> SYMLINK+="disk/by-id/md-name-$env{MD_NAME}",
> >> OPTIONS+="string_escape=replace"
> >> ENV{DEVTYPE}=="disk", ENV{MD_UUID}=="?*",
> >> SYMLINK+="disk/by-id/md-uuid-$env{MD_UUID}"
> >> ENV{DEVTYPE}=="disk", ENV{MD_DEVNAME}=="?*", TAG+="systemd",
> >> SYMLINK+="md/$env{MD_DEVNAME}"
> >> ENV{DEVTYPE}=="partition", ENV{MD_NAME}=="?*",
> >> SYMLINK+="disk/by-id/md-name-$env{MD_NAME}-part%n",
> >> OPTIONS+="string_escape=replace"
> >> ENV{DEVTYPE}=="partition", ENV{MD_UUID}=="?*",
> >> SYMLINK+="disk/by-id/md-uuid-$env{MD_UUID}-part%n"
> >> ENV{DEVTYPE}=="partition", ENV{MD_DEVNAME}=="*[^0-9]",
> >> SYMLINK+="md/$env{MD_DEVNAME}%n"
> >> ENV{DEVTYPE}=="partition", ENV{MD_DEVNAME}=="*[0-9]",
> >> SYMLINK+="md/$env{MD_DEVNAME}p%n"
> >>
> >>
> >> IMPORT{builtin}="blkid"
> >> OPTIONS+="link_priority=100"
> >> OPTIONS+="watch"
> >> ENV{ID_FS_USAGE}=="filesystem|other|crypto", ENV{ID_FS_UUID_ENC}=="?*",
> >> SYMLINK+="disk/by-uuid/$env{ID_FS_UUID_ENC}"
> >> ENV{ID_FS_USAGE}=="filesystem|other", ENV{ID_PART_ENTRY_UUID}=="?*",
> >> SYMLINK+="disk/by-partuuid/$env{ID_PART_ENTRY_UUID}"
> >> ENV{ID_FS_USAGE}=="filesystem|other", ENV{ID_FS_LABEL_ENC}=="?*",
> >> SYMLINK+="disk/by-label/$env{ID_FS_LABEL_ENC}"
> >>
> >> ENV{MD_LEVEL}=="raid[1-9]*", ENV{SYSTEMD_WANTS}+="mdmonitor.service"
> >> ENV{MD_LEVEL}=="raid[1-9]*", ENV{SYSTEMD_WANTS}+="hello.service"
> >>
> >> #RUN{program}+="/usr/bin/echo SYSTEMD_READY = $env(SYSTEMD_READY)"
> >> RUN{program}+="/usr/bin/udevadm info --query=property --name=/dev/md127"
> >>
> >> # Tell systemd to run mdmon for our container, if we need it.
> >> ENV{MD_LEVEL}=="raid[1-9]*", ENV{MD_CONTAINER}=="?*",
> >> PROGRAM="/usr/bin/readlink $env{MD_CONTAINER}", ENV{MD_MON_THIS}="%c"
> >> ENV{MD_MON_THIS}=="?*", PROGRAM="/usr/bin/basename $env{MD_MON_THIS}",
> >> ENV{SYSTEMD_WANTS}+="mdmon@%c.service"
> >> ENV{RESHAPE_ACTIVE}=="yes", PROGRAM="/usr/bin/basename
> >> $env{MD_MON_THIS}", ENV{SYSTEMD_WANTS}+="mdadm-grow-continue@%c.service"
> >>
> >> LABEL="md_end"
> >>
> >>
> >>
> Please see my second message, where I think this is because of a
> conflicting rule in 69-dm-lvm.rules:49 which later resets it:
>
> ENV{LVM_MD_PV_ACTIVATED}!="1", ENV{SYSTEMD_READY}="0"
>

I'm not familiar with this so I can't give my answer. Sorry for this.
I just right
did some jobs related with udev rules and tried this debug method.

If you are suspicious about if SYSTEMD_READY is changed or not, maybe
you can add some debug logs before and after the suspicious place?

Regards
Xiao





[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux