Re: slowness with kernel 6.4.10 and software raid

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



The above makes it very clear what is happening.   What kind of disks
are these?  And did you set the scterc timeout?  You can see it via
smartctl -l scterc /dev/sda   and then repeat on the other disk.

Setting the timeout as low as you can will improve this situation
some, but it appears that sda has a number of bad sectors on it.

a full output of "smartclt --xall /dev/sda" would be useful also to
see how bad it is.

Short answer is you probably need a new device for sda.

On Fri, Aug 18, 2023 at 1:30 PM Ranjan Maitra <mlmaitra@xxxxxxx> wrote:
>
> Thanks, Roger!
>
>
> On Fri Aug18'23 12:23:23PM, Roger Heflin wrote:
> > From: Roger Heflin <rogerheflin@xxxxxxxxx>
> > Date: Fri, 18 Aug 2023 12:23:23 -0500
> > To: Community support for Fedora users <users@xxxxxxxxxxxxxxxxxxxxxxx>
> > Reply-To: Community support for Fedora users <users@xxxxxxxxxxxxxxxxxxxxxxx>
> > Subject: Re: slowness with kernel 6.4.10 and software raid
> >
> > Is it moving at all or just stopped?  If just stopped it appears that
> > md126 is using external:/md127 for something and md127 looks wrong
> > (both disks are spare) but I don't know in this external case what
> > md127 should look like.
>
> It is moving, slowly. It is a 2 TB drive, but this is weird.
>
> >
> > I would suggest checking messages with grep md12[67] /var/log/messages
> > (and older messages files if the reboot was not this week) to see what
> > is going on.
>
> Good idea! Here is the result from
>
> $ grep md126  /var/log/messages
>
>
>   Aug 14 15:02:30 localhost mdadm[1035]: Rebuild60 event detected on md device /dev/md126
>   Aug 16 14:21:20 localhost kernel: md/raid1:md126: active with 2 out of 2 mirrors
>   Aug 16 14:21:20 localhost kernel: md126: detected capacity change from 0 to 3711741952
>   Aug 16 14:21:20 localhost kernel: md126: p1
>   Aug 16 14:21:23 localhost systemd[1]: Condition check resulted in dev-md126p1.device - /dev/md126p1 being skipped.
>   Aug 16 14:21:28 localhost systemd-fsck[942]: /dev/md126p1: clean, 7345384/115998720 files, 409971205/463967488 blocks
>   Aug 16 14:21:31 localhost kernel: EXT4-fs (md126p1): mounted filesystem 932eb81c-2ab4-4e6e-b093-46e43dbd6c28 r/w with ordered data mode. Quota mode: none.
>   Aug 16 14:21:31 localhost mdadm[1033]: NewArray event detected on md device /dev/md126
>   Aug 16 14:21:31 localhost mdadm[1033]: RebuildStarted event detected on md device /dev/md126
>   Aug 16 14:21:31 localhost kernel: md: data-check of RAID array md126
>   Aug 16 19:33:18 localhost kernel: md/raid1:md126: sda: rescheduling sector 2735900352
>   Aug 16 19:33:22 localhost kernel: md/raid1:md126: sda: rescheduling sector 2735900864
>   Aug 16 19:33:28 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735900496 on sda)
>   Aug 16 19:33:36 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735900568 on sda)
>   Aug 16 19:33:41 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735900576 on sda)
>   Aug 16 19:33:50 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735900624 on sda)
>   Aug 16 19:34:00 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735900640 on sda)
>   Aug 16 19:34:10 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735900688 on sda)
>   Aug 16 19:34:18 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735900712 on sda)
>   Aug 16 19:34:28 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735900792 on sda)
>   Aug 16 19:34:32 localhost kernel: md/raid1:md126: redirecting sector 2735900352 to other mirror: sdc
>   Aug 16 19:34:37 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735900872 on sda)
>   Aug 16 19:34:45 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735900920 on sda)
>   Aug 16 19:34:54 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735900992 on sda)
>   Aug 16 19:34:54 localhost kernel: md/raid1:md126: redirecting sector 2735900864 to other mirror: sdc
>   Aug 16 19:35:07 localhost kernel: md/raid1:md126: sda: rescheduling sector 2735905704
>   Aug 16 19:35:11 localhost kernel: md/raid1:md126: sda: rescheduling sector 2735905960
>   Aug 16 19:35:18 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735905768 on sda)
>   Aug 16 19:35:19 localhost kernel: md/raid1:md126: redirecting sector 2735905704 to other mirror: sdc
>   Aug 16 19:35:24 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735906120 on sda)
>   Aug 16 19:35:33 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735906192 on sda)
>   Aug 16 19:35:39 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735906448 on sda)
>   Aug 16 19:35:40 localhost kernel: md/raid1:md126: redirecting sector 2735905960 to other mirror: sdc
>   Aug 16 19:35:45 localhost kernel: md/raid1:md126: sda: rescheduling sector 2735906472
>   Aug 16 19:35:49 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735906504 on sda)
>   Aug 16 19:35:52 localhost kernel: md/raid1:md126: redirecting sector 2735906472 to other mirror: sdc
>   Aug 16 19:36:03 localhost kernel: md/raid1:md126: sda: rescheduling sector 2735908008
>   Aug 16 19:36:08 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735908232 on sda)
>   Aug 16 19:36:16 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735908344 on sda)
>   Aug 16 19:36:21 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735908424 on sda)
>   Aug 16 19:36:21 localhost kernel: md/raid1:md126: redirecting sector 2735908008 to other mirror: sda
>   Aug 16 19:36:30 localhost kernel: md/raid1:md126: sda: rescheduling sector 2735908008
>   Aug 16 19:36:37 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735908296 on sda)
>   Aug 16 19:36:38 localhost kernel: md/raid1:md126: redirecting sector 2735908008 to other mirror: sdc
>   Aug 16 19:36:42 localhost kernel: md/raid1:md126: sda: rescheduling sector 2735908776
>   Aug 16 19:36:42 localhost kernel: md/raid1:md126: sda: rescheduling sector 2735909032
>   Aug 16 19:36:46 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735908784 on sda)
>   Aug 16 19:36:50 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735908944 on sda)
>   Aug 16 19:36:50 localhost kernel: md/raid1:md126: redirecting sector 2735908776 to other mirror: sdc
>   Aug 16 19:36:55 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735909312 on sda)
>   Aug 16 19:37:00 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735909360 on sda)
>   Aug 16 19:37:04 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735909400 on sda)
>   Aug 16 19:37:11 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735909520 on sda)
>   Aug 16 19:37:11 localhost kernel: md/raid1:md126: redirecting sector 2735909032 to other mirror: sdc
>   Aug 16 19:37:21 localhost kernel: md/raid1:md126: sda: rescheduling sector 2735910056
>   Aug 16 19:37:21 localhost kernel: md/raid1:md126: sda: rescheduling sector 2735910568
>   Aug 16 19:37:25 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735910064 on sda)
>   Aug 16 19:37:31 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735910080 on sda)
>   Aug 16 19:38:00 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735910128 on sda)
>   Aug 16 19:38:08 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735910240 on sda)
>   Aug 16 19:38:12 localhost kernel: md/raid1:md126: redirecting sector 2735910056 to other mirror: sdc
>   Aug 16 19:38:15 localhost kernel: md/raid1:md126: redirecting sector 2735910568 to other mirror: sdc
>   Aug 16 19:38:23 localhost kernel: md/raid1:md126: sda: rescheduling sector 2735911080
>   Aug 16 19:38:23 localhost kernel: md/raid1:md126: sda: rescheduling sector 2735911592
>   Aug 16 19:38:27 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735911520 on sda)
>   Aug 16 19:38:27 localhost kernel: md/raid1:md126: redirecting sector 2735911080 to other mirror: sdc
>   Aug 16 19:38:28 localhost kernel: md/raid1:md126: redirecting sector 2735911592 to other mirror: sdc
>   Aug 16 19:38:33 localhost kernel: md/raid1:md126: sda: rescheduling sector 2735912104
>   Aug 16 19:38:37 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735912184 on sda)
>   Aug 16 19:38:45 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735912240 on sda)
>   Aug 16 19:38:49 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735912248 on sda)
>   Aug 16 19:38:59 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735912288 on sda)
>   Aug 16 19:39:05 localhost kernel: md/raid1:md126: redirecting sector 2735912104 to other mirror: sdc
>   Aug 16 19:39:10 localhost kernel: md/raid1:md126: sda: rescheduling sector 2735912872
>   Aug 16 19:39:14 localhost kernel: md/raid1:md126: sda: rescheduling sector 2735913128
>   Aug 16 19:39:25 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735912976 on sda)
>   Aug 16 19:39:33 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735913048 on sda)
>   Aug 16 19:39:37 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735913072 on sda)
>   Aug 16 19:39:41 localhost kernel: md/raid1:md126: redirecting sector 2735912872 to other mirror: sdc
>   Aug 16 19:39:45 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735913128 on sda)
>   Aug 16 19:39:55 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735913176 on sda)
>   Aug 16 19:40:05 localhost kernel: md/raid1:md126: read error corrected (8 sectors at 2735913232 on sda)
>
>
> And here is what I get from:
>
> $ grep  md127  /var/log/messages
>
>
>   Aug 16 14:16:38 localhost systemd[1]: mdmon@md127.service: Deactivated successfully.
>   Aug 16 14:16:38 localhost systemd[1]: mdmon@md127.service: Unit process 884 (mdmon) remains running after unit stopped.
>   Aug 16 14:16:38 localhost systemd[1]: Stopped mdmon@md127.service - MD Metadata Monitor on /dev/md127.
>   Aug 16 14:16:38 localhost audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=mdmon@md127 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
>   Aug 16 14:16:38 localhost systemd[1]: mdmon@md127.service: Consumed 41.719s CPU time.
>   Aug 16 14:21:20 localhost systemd[1]: Starting mdmon@md127.service - MD Metadata Monitor on /dev/md127...
>   Aug 16 14:21:20 localhost systemd[1]: Started mdmon@md127.service - MD Metadata Monitor on /dev/md127.
>   Aug 16 14:21:20 localhost audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=mdmon@md127 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
>
> >
> > Maybe also if you have a prior good reboot in messages file include
> > that and see what happened differently between the 2.
>
> Yeah, I do not know where to find this. I looked into /var/log/messages, but it looks like it starts on August 13, which was a surprise to me, and the last non-responsive instance for me was last week (August 10, I think, when I booted into the 6.4 kernel). I did reboot in frustration on August 16.
>
> Thanks,
> Ranjan
>
>
> >
> > On Fri, Aug 18, 2023 at 7:46 AM Ranjan Maitra <mlmaitra@xxxxxxx> wrote:
> > >
> > > On Thu Aug17'23 10:37:29PM, Samuel Sieb wrote:
> > > > From: Samuel Sieb <samuel@xxxxxxxx>
> > > > Date: Thu, 17 Aug 2023 22:37:29 -0700
> > > > To: users@xxxxxxxxxxxxxxxxxxxxxxx
> > > > Reply-To: Community support for Fedora users <users@xxxxxxxxxxxxxxxxxxxxxxx>
> > > > Subject: Re: slowness with kernel 6.4.10 and software raid
> > > >
> > > > On 8/17/23 21:38, Ranjan Maitra wrote:
> > > > > $ cat /proc/mdstat
> > > > >   Personalities : [raid1]
> > > > >   md126 : active raid1 sda[1] sdc[0]
> > > > >         1855870976 blocks super external:/md127/0 [2/2] [UU]
> > > > >         [=>...................]  check =  8.8% (165001216/1855870976) finish=45465.2min speed=619K/sec
> > > > >
> > > > >   md127 : inactive sda[1](S) sdc[0](S)
> > > > >         10402 blocks super external:imsm
> > > > >
> > > > >   unused devices: <none>
> > > > >
> > > > > I am not sure what it is doing, and I am a bit concerned that this will go on at this rate for about 20 days. No knowing what will happen after that, and also if this problem will recur with another reboot.
> > > >
> > > > After a certain amount of time, mdraid will do a verification of the data
> > > > where it scans the entire array.  If you reboot, it will continue from where
> > > > it left off.  But that is *really* slow, so you should find out what's going
> > > > on there.
> > >
> > > Yes, I know, just not sure what to do. Thanks very much!
> > >
> > > Any suggestion is appreciated!
> > >
> > > Best wishes,
> > > Ranjan
> > > _______________________________________________
> > > users mailing list -- users@xxxxxxxxxxxxxxxxxxxxxxx
> > > To unsubscribe send an email to users-leave@xxxxxxxxxxxxxxxxxxxxxxx
> > > Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/
> > > List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
> > > List Archives: https://lists.fedoraproject.org/archives/list/users@xxxxxxxxxxxxxxxxxxxxxxx
> > > Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
> > _______________________________________________
> > users mailing list -- users@xxxxxxxxxxxxxxxxxxxxxxx
> > To unsubscribe send an email to users-leave@xxxxxxxxxxxxxxxxxxxxxxx
> > Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/
> > List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
> > List Archives: https://lists.fedoraproject.org/archives/list/users@xxxxxxxxxxxxxxxxxxxxxxx
> > Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
> _______________________________________________
> users mailing list -- users@xxxxxxxxxxxxxxxxxxxxxxx
> To unsubscribe send an email to users-leave@xxxxxxxxxxxxxxxxxxxxxxx
> Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/
> List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
> List Archives: https://lists.fedoraproject.org/archives/list/users@xxxxxxxxxxxxxxxxxxxxxxx
> Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
_______________________________________________
users mailing list -- users@xxxxxxxxxxxxxxxxxxxxxxx
To unsubscribe send an email to users-leave@xxxxxxxxxxxxxxxxxxxxxxx
Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: https://lists.fedoraproject.org/archives/list/users@xxxxxxxxxxxxxxxxxxxxxxx
Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue



[Index of Archives]     [Older Fedora Users]     [Fedora Announce]     [Fedora Package Announce]     [EPEL Announce]     [EPEL Devel]     [Fedora Magazine]     [Fedora Summer Coding]     [Fedora Laptop]     [Fedora Cloud]     [Fedora Advisory Board]     [Fedora Education]     [Fedora Security]     [Fedora Scitech]     [Fedora Robotics]     [Fedora Infrastructure]     [Fedora Websites]     [Anaconda Devel]     [Fedora Devel Java]     [Fedora Desktop]     [Fedora Fonts]     [Fedora Marketing]     [Fedora Management Tools]     [Fedora Mentors]     [Fedora Package Review]     [Fedora R Devel]     [Fedora PHP Devel]     [Kickstart]     [Fedora Music]     [Fedora Packaging]     [Fedora SELinux]     [Fedora Legal]     [Fedora Kernel]     [Fedora OCaml]     [Coolkey]     [Virtualization Tools]     [ET Management Tools]     [Yum Users]     [Yosemite News]     [Gnome Users]     [KDE Users]     [Fedora Art]     [Fedora Docs]     [Fedora Sparc]     [Libvirt Users]     [Fedora ARM]

  Powered by Linux