Hi, On Wed, Sep 6, 2023 at 11:13 PM Yu Kuai <yukuai1@xxxxxxxxxxxxxxx> wrote: > > Hi, > > 在 2023/09/07 13:44, Jason Moss 写道: > > Hi, > > > > On Wed, Sep 6, 2023 at 6:38 PM Yu Kuai <yukuai1@xxxxxxxxxxxxxxx> wrote: > >> > >> Hi, > >> > >> 在 2023/09/06 22:05, Jason Moss 写道: > >>> Hi Kuai, > >>> > >>> I ended up using gdb rather than addr2line, as that output didn't give > >>> me the global offset. Maybe there's a better way, but this seems to be > >>> similar to what I expected. > >> > >> It's ok. > >>> > >>> (gdb) list *(reshape_request+0x416) > >>> 0x11566 is in reshape_request (drivers/md/raid5.c:6396). > >>> 6391 if ((mddev->reshape_backwards > >>> 6392 ? (safepos > writepos && readpos < writepos) > >>> 6393 : (safepos < writepos && readpos > writepos)) || > >>> 6394 time_after(jiffies, conf->reshape_checkpoint + 10*HZ)) { > >>> 6395 /* Cannot proceed until we've updated the > >>> superblock... */ > >>> 6396 wait_event(conf->wait_for_overlap, > >>> 6397 atomic_read(&conf->reshape_stripes)==0 > >>> 6398 || test_bit(MD_RECOVERY_INTR, > >> > >> If reshape is stuck here, which means: > >> > >> 1) Either reshape io is stuck somewhere and never complete; > >> 2) Or the counter reshape_stripes is broken; > >> > >> Can you read following debugfs files to verify if io is stuck in > >> underlying disk? > >> > >> /sys/kernel/debug/block/[disk]/hctx*/{sched_tags,tags,busy,dispatch} > >> > > > > I'll attach this below. > > > >> Furthermore, echo frozen should break above wait_event() because > >> 'MD_RECOVERY_INTR' will be set, however, based on your description, > >> the problem still exist. Can you collect stack and addr2line result > >> of stuck thread after echo frozen? > >> > > > > I echo'd frozen to /sys/block/md0/md/sync_action, however the echo > > call has been sitting for about 30 minutes, maybe longer, and has not > > returned. Here's the current state: > > > > root 454 0.0 0.0 0 0 ? I< Sep05 0:00 [raid5wq] > > root 455 0.0 0.0 34680 5988 ? D Sep05 0:00 (udev-worker) > > Can you also show the stack of udev-worker? And any other thread with > 'D' state, I think above "echo frozen" is probably also stuck in D > state. > As requested: ps aux | grep D USER PID %CPU %MEM VSZ RSS TTY STAT START TIME COMMAND root 455 0.0 0.0 34680 5988 ? D Sep05 0:00 (udev-worker) root 457 0.0 0.0 0 0 ? D Sep05 0:00 [md0_reshape] root 45507 0.0 0.0 8272 4736 pts/1 Ds+ Sep05 0:00 -bash jason 279169 0.0 0.0 6976 2560 pts/0 S+ 23:16 0:00 grep --color=auto D [jason@arch md]$ sudo cat /proc/455/stack [<0>] wait_woken+0x54/0x60 [<0>] raid5_make_request+0x5fe/0x12f0 [raid456] [<0>] md_handle_request+0x135/0x220 [md_mod] [<0>] __submit_bio+0xb3/0x170 [<0>] submit_bio_noacct_nocheck+0x159/0x370 [<0>] block_read_full_folio+0x21c/0x340 [<0>] filemap_read_folio+0x40/0xd0 [<0>] filemap_get_pages+0x475/0x630 [<0>] filemap_read+0xd9/0x350 [<0>] blkdev_read_iter+0x6b/0x1b0 [<0>] vfs_read+0x201/0x350 [<0>] ksys_read+0x6f/0xf0 [<0>] do_syscall_64+0x60/0x90 [<0>] entry_SYSCALL_64_after_hwframe+0x6e/0xd8 [jason@arch md]$ sudo cat /proc/45507/stack [<0>] kthread_stop+0x6a/0x180 [<0>] md_unregister_thread+0x29/0x60 [md_mod] [<0>] action_store+0x168/0x320 [md_mod] [<0>] md_attr_store+0x86/0xf0 [md_mod] [<0>] kernfs_fop_write_iter+0x136/0x1d0 [<0>] vfs_write+0x23e/0x420 [<0>] ksys_write+0x6f/0xf0 [<0>] do_syscall_64+0x60/0x90 [<0>] entry_SYSCALL_64_after_hwframe+0x6e/0xd8 Please let me know if you'd like me to identify the lines for any of those. Thanks, Jason > > root 456 99.9 0.0 0 0 ? R Sep05 1543:40 [md0_raid6] > > root 457 0.0 0.0 0 0 ? D Sep05 0:00 [md0_reshape] > > > > [jason@arch md]$ sudo cat /proc/457/stack > > [<0>] md_do_sync+0xef2/0x11d0 [md_mod] > > [<0>] md_thread+0xae/0x190 [md_mod] > > [<0>] kthread+0xe8/0x120 > > [<0>] ret_from_fork+0x34/0x50 > > [<0>] ret_from_fork_asm+0x1b/0x30 > > > > Reading symbols from md-mod.ko... > > (gdb) list *(md_do_sync+0xef2) > > 0xb3a2 is in md_do_sync (drivers/md/md.c:9035). > > 9030 ? "interrupted" : "done"); > > 9031 /* > > 9032 * this also signals 'finished resyncing' to md_stop > > 9033 */ > > 9034 blk_finish_plug(&plug); > > 9035 wait_event(mddev->recovery_wait, > > !atomic_read(&mddev->recovery_active)); > > That's also wait for reshape io to be done from common layer. > > > 9036 > > 9037 if (!test_bit(MD_RECOVERY_RESHAPE, &mddev->recovery) && > > 9038 !test_bit(MD_RECOVERY_INTR, &mddev->recovery) && > > 9039 mddev->curr_resync >= MD_RESYNC_ACTIVE) { > > > > > > The debugfs info: > > > > [root@arch ~]# cat > > /sys/kernel/debug/block/sda/hctx*/{sched_tags,tags,busy,dispatch} > > Only sched_tags is read, sorry that I didn't mean to use this exact cmd. > > Perhaps you can using following cmd: > > find /sys/kernel/debug/block/sda/ -type f | xargs grep . > > > nr_tags=64 > > nr_reserved_tags=0 > > active_queues=0 > > > > bitmap_tags: > > depth=64 > > busy=1 > > This means there is one IO in sda, however, I need more information to > make sure where is this IO. And please make sure don't run any other > thread that can read/write from sda. You can use "iostat -dmx 1" and > observe for a while to confirm that there is no new io. > > Thanks, > Kuai > > > cleared=55 > > bits_per_word=16 > > map_nr=4 > > alloc_hint={40, 20, 46, 0} > > wake_batch=8 > > wake_index=0 > > ws_active=0 > > ws={ > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > } > > round_robin=1 > > min_shallow_depth=48 > > nr_tags=32 > > nr_reserved_tags=0 > > active_queues=0 > > > > bitmap_tags: > > depth=32 > > busy=0 > > cleared=27 > > bits_per_word=8 > > map_nr=4 > > alloc_hint={19, 26, 5, 21} > > wake_batch=4 > > wake_index=0 > > ws_active=0 > > ws={ > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > } > > round_robin=1 > > min_shallow_depth=4294967295 > > > > > > > > [root@arch ~]# cat /sys/kernel/debug/block/sdb/hctx* > > /{sched_tags,tags,busy,dispatch} > > nr_tags=64 > > nr_reserved_tags=0 > > active_queues=0 > > > > bitmap_tags: > > depth=64 > > busy=1 > > cleared=56 > > bits_per_word=16 > > map_nr=4 > > alloc_hint={57, 43, 14, 19} > > wake_batch=8 > > wake_index=0 > > ws_active=0 > > ws={ > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > } > > round_robin=1 > > min_shallow_depth=48 > > nr_tags=32 > > nr_reserved_tags=0 > > active_queues=0 > > > > bitmap_tags: > > depth=32 > > busy=0 > > cleared=24 > > bits_per_word=8 > > map_nr=4 > > alloc_hint={17, 13, 23, 17} > > wake_batch=4 > > wake_index=0 > > ws_active=0 > > ws={ > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > } > > round_robin=1 > > min_shallow_depth=4294967295 > > > > > > [root@arch ~]# cat > > /sys/kernel/debug/block/sdd/hctx*/{sched_tags,tags,busy,dispatch} > > nr_tags=64 > > nr_reserved_tags=0 > > active_queues=0 > > > > bitmap_tags: > > depth=64 > > busy=1 > > cleared=51 > > bits_per_word=16 > > map_nr=4 > > alloc_hint={36, 43, 15, 7} > > wake_batch=8 > > wake_index=0 > > ws_active=0 > > ws={ > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > } > > round_robin=1 > > min_shallow_depth=48 > > nr_tags=32 > > nr_reserved_tags=0 > > active_queues=0 > > > > bitmap_tags: > > depth=32 > > busy=0 > > cleared=31 > > bits_per_word=8 > > map_nr=4 > > alloc_hint={0, 15, 1, 22} > > wake_batch=4 > > wake_index=0 > > ws_active=0 > > ws={ > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > } > > round_robin=1 > > min_shallow_depth=4294967295 > > > > > > [root@arch ~]# cat > > /sys/kernel/debug/block/sdf/hctx*/{sched_tags,tags,busy,dispatch} > > nr_tags=256 > > nr_reserved_tags=0 > > active_queues=0 > > > > bitmap_tags: > > depth=256 > > busy=1 > > cleared=131 > > bits_per_word=64 > > map_nr=4 > > alloc_hint={125, 46, 83, 205} > > wake_batch=8 > > wake_index=0 > > ws_active=0 > > ws={ > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > } > > round_robin=0 > > min_shallow_depth=192 > > nr_tags=10104 > > nr_reserved_tags=0 > > active_queues=0 > > > > bitmap_tags: > > depth=10104 > > busy=0 > > cleared=235 > > bits_per_word=64 > > map_nr=158 > > alloc_hint={503, 2913, 9827, 9851} > > wake_batch=8 > > wake_index=0 > > ws_active=0 > > ws={ > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > } > > round_robin=0 > > min_shallow_depth=4294967295 > > > > > > [root@arch ~]# cat > > /sys/kernel/debug/block/sdh/hctx*/{sched_tags,tags,busy,dispatch} > > nr_tags=256 > > nr_reserved_tags=0 > > active_queues=0 > > > > bitmap_tags: > > depth=256 > > busy=1 > > cleared=97 > > bits_per_word=64 > > map_nr=4 > > alloc_hint={144, 144, 127, 254} > > wake_batch=8 > > wake_index=0 > > ws_active=0 > > ws={ > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > } > > round_robin=0 > > min_shallow_depth=192 > > nr_tags=10104 > > nr_reserved_tags=0 > > active_queues=0 > > > > bitmap_tags: > > depth=10104 > > busy=0 > > cleared=235 > > bits_per_word=64 > > map_nr=158 > > alloc_hint={503, 2913, 9827, 9851} > > wake_batch=8 > > wake_index=0 > > ws_active=0 > > ws={ > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > } > > round_robin=0 > > min_shallow_depth=4294967295 > > > > > > [root@arch ~]# cat > > /sys/kernel/debug/block/sdi/hctx*/{sched_tags,tags,busy,dispatch} > > nr_tags=256 > > nr_reserved_tags=0 > > active_queues=0 > > > > bitmap_tags: > > depth=256 > > busy=1 > > cleared=34 > > bits_per_word=64 > > map_nr=4 > > alloc_hint={197, 20, 1, 230} > > wake_batch=8 > > wake_index=0 > > ws_active=0 > > ws={ > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > } > > round_robin=0 > > min_shallow_depth=192 > > nr_tags=10104 > > nr_reserved_tags=0 > > active_queues=0 > > > > bitmap_tags: > > depth=10104 > > busy=0 > > cleared=235 > > bits_per_word=64 > > map_nr=158 > > alloc_hint={503, 2913, 9827, 9851} > > wake_batch=8 > > wake_index=0 > > ws_active=0 > > ws={ > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > } > > round_robin=0 > > min_shallow_depth=4294967295 > > > > > > [root@arch ~]# cat > > /sys/kernel/debug/block/sdj/hctx*/{sched_tags,tags,busy,dispatch} > > nr_tags=256 > > nr_reserved_tags=0 > > active_queues=0 > > > > bitmap_tags: > > depth=256 > > busy=1 > > cleared=27 > > bits_per_word=64 > > map_nr=4 > > alloc_hint={132, 74, 129, 76} > > wake_batch=8 > > wake_index=0 > > ws_active=0 > > ws={ > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > } > > round_robin=0 > > min_shallow_depth=192 > > nr_tags=10104 > > nr_reserved_tags=0 > > active_queues=0 > > > > bitmap_tags: > > depth=10104 > > busy=0 > > cleared=235 > > bits_per_word=64 > > map_nr=158 > > alloc_hint={503, 2913, 9827, 9851} > > wake_batch=8 > > wake_index=0 > > ws_active=0 > > ws={ > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > {.wait=inactive}, > > } > > round_robin=0 > > min_shallow_depth=4294967295 > > > > > > Thanks for your continued assistance with this! > > Jason > > > > > >> Thanks, > >> Kuai > >> > >>> &mddev->recovery)); > >>> 6399 if (atomic_read(&conf->reshape_stripes) != 0) > >>> 6400 return 0; > >>> > >>> Thanks > >>> > >>> On Mon, Sep 4, 2023 at 6:08 PM Yu Kuai <yukuai1@xxxxxxxxxxxxxxx> wrote: > >>>> > >>>> Hi, > >>>> > >>>> 在 2023/09/05 0:38, Jason Moss 写道: > >>>>> Hi Kuai, > >>>>> > >>>>> Thank you for the suggestion, I was previously on 5.15.0. I've built > >>>>> an environment with 6.5.0.1 now and assembled the array there, but the > >>>>> same problem happens. It reshaped for 20-30 seconds, then completely > >>>>> stopped. > >>>>> > >>>>> Processes and /proc/<PID>/stack output: > >>>>> root 24593 0.0 0.0 0 0 ? I< 09:22 0:00 [raid5wq] > >>>>> root 24594 96.5 0.0 0 0 ? R 09:22 2:29 [md0_raid6] > >>>>> root 24595 0.3 0.0 0 0 ? D 09:22 0:00 [md0_reshape] > >>>>> > >>>>> [root@arch ~]# cat /proc/24593/stack > >>>>> [<0>] rescuer_thread+0x2b0/0x3b0 > >>>>> [<0>] kthread+0xe8/0x120 > >>>>> [<0>] ret_from_fork+0x34/0x50 > >>>>> [<0>] ret_from_fork_asm+0x1b/0x30 > >>>>> > >>>>> [root@arch ~]# cat /proc/24594/stack > >>>>> > >>>>> [root@arch ~]# cat /proc/24595/stack > >>>>> [<0>] reshape_request+0x416/0x9f0 [raid456] > >>>> Can you provide the addr2line result? Let's see where reshape_request() > >>>> is stuck first. > >>>> > >>>> Thanks, > >>>> Kuai > >>>> > >>>>> [<0>] raid5_sync_request+0x2fc/0x3d0 [raid456] > >>>>> [<0>] md_do_sync+0x7d6/0x11d0 [md_mod] > >>>>> [<0>] md_thread+0xae/0x190 [md_mod] > >>>>> [<0>] kthread+0xe8/0x120 > >>>>> [<0>] ret_from_fork+0x34/0x50 > >>>>> [<0>] ret_from_fork_asm+0x1b/0x30 > >>>>> > >>>>> Please let me know if there's a better way to provide the stack info. > >>>>> > >>>>> Thank you > >>>>> > >>>>> On Sun, Sep 3, 2023 at 6:41 PM Yu Kuai <yukuai1@xxxxxxxxxxxxxxx> wrote: > >>>>>> > >>>>>> Hi, > >>>>>> > >>>>>> 在 2023/09/04 5:39, Jason Moss 写道: > >>>>>>> Hello, > >>>>>>> > >>>>>>> I recently attempted to add a new drive to my 8-drive RAID 6 array, > >>>>>>> growing it to 9 drives. I've done similar before with the same array, > >>>>>>> having previously grown it from 6 drives to 7 and then from 7 to 8 > >>>>>>> with no issues. Drives are WD Reds, most older than 2019, some > >>>>>>> (including the newest) newer, but all confirmed CMR and not SMR. > >>>>>>> > >>>>>>> Process used to expand the array: > >>>>>>> mdadm --add /dev/md0 /dev/sdb1 > >>>>>>> mdadm --grow --raid-devices=9 --backup-file=/root/grow_md0.bak /dev/md0 > >>>>>>> > >>>>>>> The reshape started off fine, the process was underway, and the volume > >>>>>>> was still usable as expected. However, 15-30 minutes into the reshape, > >>>>>>> I lost access to the contents of the drive. Checking /proc/mdstat, the > >>>>>>> reshape was stopped at 0.6% with the counter not incrementing at all. > >>>>>>> Any process accessing the array would just hang until killed. I waited > >>>>>> > >>>>>> What kernel version are you using? And it'll be very helpful if you can > >>>>>> collect the stack of all stuck thread. There is a known deadlock for > >>>>>> raid5 related to reshape, and it's fixed in v6.5: > >>>>>> > >>>>>> https://lore.kernel.org/r/20230512015610.821290-6-yukuai1@xxxxxxxxxxxxxxx > >>>>>> > >>>>>>> a half hour and there was still no further change to the counter. At > >>>>>>> this point, I restarted the server and found that when it came back up > >>>>>>> it would begin reshaping again, but only very briefly, under 30 > >>>>>>> seconds, but the counter would be increasing during that time. > >>>>>>> > >>>>>>> I searched furiously for ideas and tried stopping and reassembling the > >>>>>>> array, assembling with an invalid-backup flag, echoing "frozen" then > >>>>>>> "reshape" to the sync_action file, and echoing "max" to the sync_max > >>>>>>> file. Nothing ever seemed to make a difference. > >>>>>>> > >>>>>> > >>>>>> Don't do this before v6.5, echo "reshape" while reshape is still in > >>>>>> progress will corrupt your data: > >>>>>> > >>>>>> https://lore.kernel.org/r/20230512015610.821290-3-yukuai1@xxxxxxxxxxxxxxx > >>>>>> > >>>>>> Thanks, > >>>>>> Kuai > >>>>>> > >>>>>>> Here is where I slightly panicked, worried that I'd borked my array, > >>>>>>> and powered off the server again and disconnected the new drive that > >>>>>>> was just added, assuming that since it was the change, it may be the > >>>>>>> problem despite having burn-in tested it, and figuring that I'll rush > >>>>>>> order a new drive, so long as the reshape continues and I can just > >>>>>>> rebuild onto a new drive once the reshape finishes. However, this made > >>>>>>> no difference and the array continued to not rebuild. > >>>>>>> > >>>>>>> Much searching later, I'd found nothing substantially different then > >>>>>>> I'd already tried and one of the common threads in other people's > >>>>>>> issues was bad drives, so I ran a self-test against each of the > >>>>>>> existing drives and found one drive that failed the read test. > >>>>>>> Thinking I had the culprit now, I dropped that drive out of the array > >>>>>>> and assembled the array again, but the same behavior persists. The > >>>>>>> array reshapes very briefly, then completely stops. > >>>>>>> > >>>>>>> Down to 0 drives of redundancy (in the reshaped section at least), not > >>>>>>> finding any new ideas on any of the forums, mailing list, wiki, etc, > >>>>>>> and very frustrated, I took a break, bought all new drives to build a > >>>>>>> new array in another server and restored from a backup. However, there > >>>>>>> is still some data not captured by the most recent backup that I would > >>>>>>> like to recover, and I'd also like to solve the problem purely to > >>>>>>> understand what happened and how to recover in the future. > >>>>>>> > >>>>>>> Is there anything else I should try to recover this array, or is this > >>>>>>> a lost cause? > >>>>>>> > >>>>>>> Details requested by the wiki to follow and I'm happy to collect any > >>>>>>> further data that would assist. /dev/sdb is the new drive that was > >>>>>>> added, then disconnected. /dev/sdh is the drive that failed a > >>>>>>> self-test and was removed from the array. > >>>>>>> > >>>>>>> Thank you in advance for any help provided! > >>>>>>> > >>>>>>> > >>>>>>> $ uname -a > >>>>>>> Linux Blyth 5.15.0-76-generic #83-Ubuntu SMP Thu Jun 15 19:16:32 UTC > >>>>>>> 2023 x86_64 x86_64 x86_64 GNU/Linux > >>>>>>> > >>>>>>> $ mdadm --version > >>>>>>> mdadm - v4.2 - 2021-12-30 > >>>>>>> > >>>>>>> > >>>>>>> $ sudo smartctl -H -i -l scterc /dev/sda > >>>>>>> smartctl 7.2 2020-12-30 r5155 [x86_64-linux-5.15.0-76-generic] (local build) > >>>>>>> Copyright (C) 2002-20, Bruce Allen, Christian Franke, www.smartmontools.org > >>>>>>> > >>>>>>> === START OF INFORMATION SECTION === > >>>>>>> Model Family: Western Digital Red > >>>>>>> Device Model: WDC WD30EFRX-68EUZN0 > >>>>>>> Serial Number: WD-WCC4N7AT7R7X > >>>>>>> LU WWN Device Id: 5 0014ee 268545f93 > >>>>>>> Firmware Version: 82.00A82 > >>>>>>> User Capacity: 3,000,592,982,016 bytes [3.00 TB] > >>>>>>> Sector Sizes: 512 bytes logical, 4096 bytes physical > >>>>>>> Rotation Rate: 5400 rpm > >>>>>>> Device is: In smartctl database [for details use: -P show] > >>>>>>> ATA Version is: ACS-2 (minor revision not indicated) > >>>>>>> SATA Version is: SATA 3.0, 6.0 Gb/s (current: 6.0 Gb/s) > >>>>>>> Local Time is: Sun Sep 3 13:27:55 2023 PDT > >>>>>>> SMART support is: Available - device has SMART capability. > >>>>>>> SMART support is: Enabled > >>>>>>> > >>>>>>> === START OF READ SMART DATA SECTION === > >>>>>>> SMART overall-health self-assessment test result: PASSED > >>>>>>> > >>>>>>> SCT Error Recovery Control: > >>>>>>> Read: 70 (7.0 seconds) > >>>>>>> Write: 70 (7.0 seconds) > >>>>>>> > >>>>>>> $ sudo smartctl -H -i -l scterc /dev/sda > >>>>>>> smartctl 7.2 2020-12-30 r5155 [x86_64-linux-5.15.0-76-generic] (local build) > >>>>>>> Copyright (C) 2002-20, Bruce Allen, Christian Franke, www.smartmontools.org > >>>>>>> > >>>>>>> === START OF INFORMATION SECTION === > >>>>>>> Model Family: Western Digital Red > >>>>>>> Device Model: WDC WD30EFRX-68EUZN0 > >>>>>>> Serial Number: WD-WCC4N7AT7R7X > >>>>>>> LU WWN Device Id: 5 0014ee 268545f93 > >>>>>>> Firmware Version: 82.00A82 > >>>>>>> User Capacity: 3,000,592,982,016 bytes [3.00 TB] > >>>>>>> Sector Sizes: 512 bytes logical, 4096 bytes physical > >>>>>>> Rotation Rate: 5400 rpm > >>>>>>> Device is: In smartctl database [for details use: -P show] > >>>>>>> ATA Version is: ACS-2 (minor revision not indicated) > >>>>>>> SATA Version is: SATA 3.0, 6.0 Gb/s (current: 6.0 Gb/s) > >>>>>>> Local Time is: Sun Sep 3 13:28:16 2023 PDT > >>>>>>> SMART support is: Available - device has SMART capability. > >>>>>>> SMART support is: Enabled > >>>>>>> > >>>>>>> === START OF READ SMART DATA SECTION === > >>>>>>> SMART overall-health self-assessment test result: PASSED > >>>>>>> > >>>>>>> SCT Error Recovery Control: > >>>>>>> Read: 70 (7.0 seconds) > >>>>>>> Write: 70 (7.0 seconds) > >>>>>>> > >>>>>>> $ sudo smartctl -H -i -l scterc /dev/sdb > >>>>>>> smartctl 7.2 2020-12-30 r5155 [x86_64-linux-5.15.0-76-generic] (local build) > >>>>>>> Copyright (C) 2002-20, Bruce Allen, Christian Franke, www.smartmontools.org > >>>>>>> > >>>>>>> === START OF INFORMATION SECTION === > >>>>>>> Model Family: Western Digital Red > >>>>>>> Device Model: WDC WD30EFRX-68EUZN0 > >>>>>>> Serial Number: WD-WXG1A8UGLS42 > >>>>>>> LU WWN Device Id: 5 0014ee 2b75ef53b > >>>>>>> Firmware Version: 80.00A80 > >>>>>>> User Capacity: 3,000,592,982,016 bytes [3.00 TB] > >>>>>>> Sector Sizes: 512 bytes logical, 4096 bytes physical > >>>>>>> Rotation Rate: 5400 rpm > >>>>>>> Device is: In smartctl database [for details use: -P show] > >>>>>>> ATA Version is: ACS-2 (minor revision not indicated) > >>>>>>> SATA Version is: SATA 3.0, 6.0 Gb/s (current: 6.0 Gb/s) > >>>>>>> Local Time is: Sun Sep 3 13:28:19 2023 PDT > >>>>>>> SMART support is: Available - device has SMART capability. > >>>>>>> SMART support is: Enabled > >>>>>>> > >>>>>>> === START OF READ SMART DATA SECTION === > >>>>>>> SMART overall-health self-assessment test result: PASSED > >>>>>>> > >>>>>>> SCT Error Recovery Control: > >>>>>>> Read: 70 (7.0 seconds) > >>>>>>> Write: 70 (7.0 seconds) > >>>>>>> > >>>>>>> $ sudo smartctl -H -i -l scterc /dev/sdc > >>>>>>> smartctl 7.2 2020-12-30 r5155 [x86_64-linux-5.15.0-76-generic] (local build) > >>>>>>> Copyright (C) 2002-20, Bruce Allen, Christian Franke, www.smartmontools.org > >>>>>>> > >>>>>>> === START OF INFORMATION SECTION === > >>>>>>> Model Family: Western Digital Red > >>>>>>> Device Model: WDC WD30EFRX-68EUZN0 > >>>>>>> Serial Number: WD-WCC4N4HYL32Y > >>>>>>> LU WWN Device Id: 5 0014ee 2630752f8 > >>>>>>> Firmware Version: 82.00A82 > >>>>>>> User Capacity: 3,000,592,982,016 bytes [3.00 TB] > >>>>>>> Sector Sizes: 512 bytes logical, 4096 bytes physical > >>>>>>> Rotation Rate: 5400 rpm > >>>>>>> Device is: In smartctl database [for details use: -P show] > >>>>>>> ATA Version is: ACS-2 (minor revision not indicated) > >>>>>>> SATA Version is: SATA 3.0, 6.0 Gb/s (current: 6.0 Gb/s) > >>>>>>> Local Time is: Sun Sep 3 13:28:20 2023 PDT > >>>>>>> SMART support is: Available - device has SMART capability. > >>>>>>> SMART support is: Enabled > >>>>>>> > >>>>>>> === START OF READ SMART DATA SECTION === > >>>>>>> SMART overall-health self-assessment test result: PASSED > >>>>>>> > >>>>>>> SCT Error Recovery Control: > >>>>>>> Read: 70 (7.0 seconds) > >>>>>>> Write: 70 (7.0 seconds) > >>>>>>> > >>>>>>> $ sudo smartctl -H -i -l scterc /dev/sdd > >>>>>>> smartctl 7.2 2020-12-30 r5155 [x86_64-linux-5.15.0-76-generic] (local build) > >>>>>>> Copyright (C) 2002-20, Bruce Allen, Christian Franke, www.smartmontools.org > >>>>>>> > >>>>>>> === START OF INFORMATION SECTION === > >>>>>>> Model Family: Western Digital Red > >>>>>>> Device Model: WDC WD30EFRX-68N32N0 > >>>>>>> Serial Number: WD-WCC7K1FF6DYK > >>>>>>> LU WWN Device Id: 5 0014ee 2ba952a30 > >>>>>>> Firmware Version: 82.00A82 > >>>>>>> User Capacity: 3,000,592,982,016 bytes [3.00 TB] > >>>>>>> Sector Sizes: 512 bytes logical, 4096 bytes physical > >>>>>>> Rotation Rate: 5400 rpm > >>>>>>> Form Factor: 3.5 inches > >>>>>>> Device is: In smartctl database [for details use: -P show] > >>>>>>> ATA Version is: ACS-3 T13/2161-D revision 5 > >>>>>>> SATA Version is: SATA 3.1, 6.0 Gb/s (current: 6.0 Gb/s) > >>>>>>> Local Time is: Sun Sep 3 13:28:21 2023 PDT > >>>>>>> SMART support is: Available - device has SMART capability. > >>>>>>> SMART support is: Enabled > >>>>>>> > >>>>>>> === START OF READ SMART DATA SECTION === > >>>>>>> SMART overall-health self-assessment test result: PASSED > >>>>>>> > >>>>>>> SCT Error Recovery Control: > >>>>>>> Read: 70 (7.0 seconds) > >>>>>>> Write: 70 (7.0 seconds) > >>>>>>> > >>>>>>> $ sudo smartctl -H -i -l scterc /dev/sde > >>>>>>> smartctl 7.2 2020-12-30 r5155 [x86_64-linux-5.15.0-76-generic] (local build) > >>>>>>> Copyright (C) 2002-20, Bruce Allen, Christian Franke, www.smartmontools.org > >>>>>>> > >>>>>>> === START OF INFORMATION SECTION === > >>>>>>> Model Family: Western Digital Red > >>>>>>> Device Model: WDC WD30EFRX-68EUZN0 > >>>>>>> Serial Number: WD-WCC4N5ZHTRJF > >>>>>>> LU WWN Device Id: 5 0014ee 2b88b83bb > >>>>>>> Firmware Version: 82.00A82 > >>>>>>> User Capacity: 3,000,592,982,016 bytes [3.00 TB] > >>>>>>> Sector Sizes: 512 bytes logical, 4096 bytes physical > >>>>>>> Rotation Rate: 5400 rpm > >>>>>>> Device is: In smartctl database [for details use: -P show] > >>>>>>> ATA Version is: ACS-2 (minor revision not indicated) > >>>>>>> SATA Version is: SATA 3.0, 6.0 Gb/s (current: 6.0 Gb/s) > >>>>>>> Local Time is: Sun Sep 3 13:28:22 2023 PDT > >>>>>>> SMART support is: Available - device has SMART capability. > >>>>>>> SMART support is: Enabled > >>>>>>> > >>>>>>> === START OF READ SMART DATA SECTION === > >>>>>>> SMART overall-health self-assessment test result: PASSED > >>>>>>> > >>>>>>> SCT Error Recovery Control: > >>>>>>> Read: 70 (7.0 seconds) > >>>>>>> Write: 70 (7.0 seconds) > >>>>>>> > >>>>>>> $ sudo smartctl -H -i -l scterc /dev/sdf > >>>>>>> smartctl 7.2 2020-12-30 r5155 [x86_64-linux-5.15.0-76-generic] (local build) > >>>>>>> Copyright (C) 2002-20, Bruce Allen, Christian Franke, www.smartmontools.org > >>>>>>> > >>>>>>> === START OF INFORMATION SECTION === > >>>>>>> Model Family: Western Digital Red > >>>>>>> Device Model: WDC WD30EFRX-68AX9N0 > >>>>>>> Serial Number: WD-WMC1T3804790 > >>>>>>> LU WWN Device Id: 5 0014ee 6036b6826 > >>>>>>> Firmware Version: 80.00A80 > >>>>>>> User Capacity: 3,000,592,982,016 bytes [3.00 TB] > >>>>>>> Sector Sizes: 512 bytes logical, 4096 bytes physical > >>>>>>> Device is: In smartctl database [for details use: -P show] > >>>>>>> ATA Version is: ACS-2 (minor revision not indicated) > >>>>>>> SATA Version is: SATA 3.0, 6.0 Gb/s (current: 6.0 Gb/s) > >>>>>>> Local Time is: Sun Sep 3 13:28:23 2023 PDT > >>>>>>> SMART support is: Available - device has SMART capability. > >>>>>>> SMART support is: Enabled > >>>>>>> > >>>>>>> === START OF READ SMART DATA SECTION === > >>>>>>> SMART overall-health self-assessment test result: PASSED > >>>>>>> > >>>>>>> SCT Error Recovery Control: > >>>>>>> Read: 70 (7.0 seconds) > >>>>>>> Write: 70 (7.0 seconds) > >>>>>>> > >>>>>>> $ sudo smartctl -H -i -l scterc /dev/sdg > >>>>>>> smartctl 7.2 2020-12-30 r5155 [x86_64-linux-5.15.0-76-generic] (local build) > >>>>>>> Copyright (C) 2002-20, Bruce Allen, Christian Franke, www.smartmontools.org > >>>>>>> > >>>>>>> === START OF INFORMATION SECTION === > >>>>>>> Model Family: Western Digital Red > >>>>>>> Device Model: WDC WD30EFRX-68EUZN0 > >>>>>>> Serial Number: WD-WMC4N0H692Z9 > >>>>>>> LU WWN Device Id: 5 0014ee 65af39740 > >>>>>>> Firmware Version: 82.00A82 > >>>>>>> User Capacity: 3,000,592,982,016 bytes [3.00 TB] > >>>>>>> Sector Sizes: 512 bytes logical, 4096 bytes physical > >>>>>>> Rotation Rate: 5400 rpm > >>>>>>> Device is: In smartctl database [for details use: -P show] > >>>>>>> ATA Version is: ACS-2 (minor revision not indicated) > >>>>>>> SATA Version is: SATA 3.0, 6.0 Gb/s (current: 6.0 Gb/s) > >>>>>>> Local Time is: Sun Sep 3 13:28:24 2023 PDT > >>>>>>> SMART support is: Available - device has SMART capability. > >>>>>>> SMART support is: Enabled > >>>>>>> > >>>>>>> === START OF READ SMART DATA SECTION === > >>>>>>> SMART overall-health self-assessment test result: PASSED > >>>>>>> > >>>>>>> SCT Error Recovery Control: > >>>>>>> Read: 70 (7.0 seconds) > >>>>>>> Write: 70 (7.0 seconds) > >>>>>>> > >>>>>>> $ sudo smartctl -H -i -l scterc /dev/sdh > >>>>>>> smartctl 7.2 2020-12-30 r5155 [x86_64-linux-5.15.0-76-generic] (local build) > >>>>>>> Copyright (C) 2002-20, Bruce Allen, Christian Franke, www.smartmontools.org > >>>>>>> > >>>>>>> === START OF INFORMATION SECTION === > >>>>>>> Model Family: Western Digital Red > >>>>>>> Device Model: WDC WD30EFRX-68EUZN0 > >>>>>>> Serial Number: WD-WMC4N0K5S750 > >>>>>>> LU WWN Device Id: 5 0014ee 6b048d9ca > >>>>>>> Firmware Version: 82.00A82 > >>>>>>> User Capacity: 3,000,592,982,016 bytes [3.00 TB] > >>>>>>> Sector Sizes: 512 bytes logical, 4096 bytes physical > >>>>>>> Rotation Rate: 5400 rpm > >>>>>>> Device is: In smartctl database [for details use: -P show] > >>>>>>> ATA Version is: ACS-2 (minor revision not indicated) > >>>>>>> SATA Version is: SATA 3.0, 6.0 Gb/s (current: 6.0 Gb/s) > >>>>>>> Local Time is: Sun Sep 3 13:28:24 2023 PDT > >>>>>>> SMART support is: Available - device has SMART capability. > >>>>>>> SMART support is: Enabled > >>>>>>> > >>>>>>> === START OF READ SMART DATA SECTION === > >>>>>>> SMART overall-health self-assessment test result: PASSED > >>>>>>> > >>>>>>> SCT Error Recovery Control: > >>>>>>> Read: 70 (7.0 seconds) > >>>>>>> Write: 70 (7.0 seconds) > >>>>>>> > >>>>>>> $ sudo smartctl -H -i -l scterc /dev/sdi > >>>>>>> smartctl 7.2 2020-12-30 r5155 [x86_64-linux-5.15.0-76-generic] (local build) > >>>>>>> Copyright (C) 2002-20, Bruce Allen, Christian Franke, www.smartmontools.org > >>>>>>> > >>>>>>> === START OF INFORMATION SECTION === > >>>>>>> Model Family: Western Digital Red > >>>>>>> Device Model: WDC WD30EFRX-68AX9N0 > >>>>>>> Serial Number: WD-WMC1T1502475 > >>>>>>> LU WWN Device Id: 5 0014ee 058d2e5cb > >>>>>>> Firmware Version: 80.00A80 > >>>>>>> User Capacity: 3,000,592,982,016 bytes [3.00 TB] > >>>>>>> Sector Sizes: 512 bytes logical, 4096 bytes physical > >>>>>>> Device is: In smartctl database [for details use: -P show] > >>>>>>> ATA Version is: ACS-2 (minor revision not indicated) > >>>>>>> SATA Version is: SATA 3.0, 6.0 Gb/s (current: 6.0 Gb/s) > >>>>>>> Local Time is: Sun Sep 3 13:28:27 2023 PDT > >>>>>>> SMART support is: Available - device has SMART capability. > >>>>>>> SMART support is: Enabled > >>>>>>> > >>>>>>> === START OF READ SMART DATA SECTION === > >>>>>>> SMART overall-health self-assessment test result: PASSED > >>>>>>> > >>>>>>> SCT Error Recovery Control: > >>>>>>> Read: 70 (7.0 seconds) > >>>>>>> Write: 70 (7.0 seconds) > >>>>>>> > >>>>>>> > >>>>>>> $ sudo mdadm --examine /dev/sda > >>>>>>> /dev/sda: > >>>>>>> MBR Magic : aa55 > >>>>>>> Partition[0] : 4294967295 sectors at 1 (type ee) > >>>>>>> $ sudo mdadm --examine /dev/sda1 > >>>>>>> /dev/sda1: > >>>>>>> Magic : a92b4efc > >>>>>>> Version : 1.2 > >>>>>>> Feature Map : 0xd > >>>>>>> Array UUID : 440dc11e:079308b1:131eda79:9a74c670 > >>>>>>> Name : Blyth:0 (local to host Blyth) > >>>>>>> Creation Time : Tue Aug 4 23:47:57 2015 > >>>>>>> Raid Level : raid6 > >>>>>>> Raid Devices : 9 > >>>>>>> > >>>>>>> Avail Dev Size : 5856376832 sectors (2.73 TiB 3.00 TB) > >>>>>>> Array Size : 20497268736 KiB (19.09 TiB 20.99 TB) > >>>>>>> Used Dev Size : 5856362496 sectors (2.73 TiB 3.00 TB) > >>>>>>> Data Offset : 247808 sectors > >>>>>>> Super Offset : 8 sectors > >>>>>>> Unused Space : before=247728 sectors, after=14336 sectors > >>>>>>> State : clean > >>>>>>> Device UUID : 8ca60ad5:60d19333:11b24820:91453532 > >>>>>>> > >>>>>>> Internal Bitmap : 8 sectors from superblock > >>>>>>> Reshape pos'n : 124311040 (118.55 GiB 127.29 GB) > >>>>>>> Delta Devices : 1 (8->9) > >>>>>>> > >>>>>>> Update Time : Tue Jul 11 23:12:08 2023 > >>>>>>> Bad Block Log : 512 entries available at offset 24 sectors - bad > >>>>>>> blocks present. > >>>>>>> Checksum : b6d8f4d1 - correct > >>>>>>> Events : 181105 > >>>>>>> > >>>>>>> Layout : left-symmetric > >>>>>>> Chunk Size : 512K > >>>>>>> > >>>>>>> Device Role : Active device 7 > >>>>>>> Array State : AA.AAAAA. ('A' == active, '.' == missing, 'R' == replacing) > >>>>>>> > >>>>>>> $ sudo mdadm --examine /dev/sdb > >>>>>>> /dev/sdb: > >>>>>>> MBR Magic : aa55 > >>>>>>> Partition[0] : 4294967295 sectors at 1 (type ee) > >>>>>>> $ sudo mdadm --examine /dev/sdb1 > >>>>>>> /dev/sdb1: > >>>>>>> Magic : a92b4efc > >>>>>>> Version : 1.2 > >>>>>>> Feature Map : 0x5 > >>>>>>> Array UUID : 440dc11e:079308b1:131eda79:9a74c670 > >>>>>>> Name : Blyth:0 (local to host Blyth) > >>>>>>> Creation Time : Tue Aug 4 23:47:57 2015 > >>>>>>> Raid Level : raid6 > >>>>>>> Raid Devices : 9 > >>>>>>> > >>>>>>> Avail Dev Size : 5856376832 sectors (2.73 TiB 3.00 TB) > >>>>>>> Array Size : 20497268736 KiB (19.09 TiB 20.99 TB) > >>>>>>> Used Dev Size : 5856362496 sectors (2.73 TiB 3.00 TB) > >>>>>>> Data Offset : 247808 sectors > >>>>>>> Super Offset : 8 sectors > >>>>>>> Unused Space : before=247728 sectors, after=14336 sectors > >>>>>>> State : clean > >>>>>>> Device UUID : 386d3001:16447e43:4d2a5459:85618d11 > >>>>>>> > >>>>>>> Internal Bitmap : 8 sectors from superblock > >>>>>>> Reshape pos'n : 124207104 (118.45 GiB 127.19 GB) > >>>>>>> Delta Devices : 1 (8->9) > >>>>>>> > >>>>>>> Update Time : Tue Jul 11 00:02:59 2023 > >>>>>>> Bad Block Log : 512 entries available at offset 24 sectors > >>>>>>> Checksum : b544a39 - correct > >>>>>>> Events : 181077 > >>>>>>> > >>>>>>> Layout : left-symmetric > >>>>>>> Chunk Size : 512K > >>>>>>> > >>>>>>> Device Role : Active device 8 > >>>>>>> Array State : AAAAAAAAA ('A' == active, '.' == missing, 'R' == replacing) > >>>>>>> > >>>>>>> $ sudo mdadm --examine /dev/sdc > >>>>>>> /dev/sdc: > >>>>>>> MBR Magic : aa55 > >>>>>>> Partition[0] : 4294967295 sectors at 1 (type ee) > >>>>>>> $ sudo mdadm --examine /dev/sdc1 > >>>>>>> /dev/sdc1: > >>>>>>> Magic : a92b4efc > >>>>>>> Version : 1.2 > >>>>>>> Feature Map : 0xd > >>>>>>> Array UUID : 440dc11e:079308b1:131eda79:9a74c670 > >>>>>>> Name : Blyth:0 (local to host Blyth) > >>>>>>> Creation Time : Tue Aug 4 23:47:57 2015 > >>>>>>> Raid Level : raid6 > >>>>>>> Raid Devices : 9 > >>>>>>> > >>>>>>> Avail Dev Size : 5856376832 sectors (2.73 TiB 3.00 TB) > >>>>>>> Array Size : 20497268736 KiB (19.09 TiB 20.99 TB) > >>>>>>> Used Dev Size : 5856362496 sectors (2.73 TiB 3.00 TB) > >>>>>>> Data Offset : 247808 sectors > >>>>>>> Super Offset : 8 sectors > >>>>>>> Unused Space : before=247720 sectors, after=14336 sectors > >>>>>>> State : clean > >>>>>>> Device UUID : 1798ec4f:72c56905:4e74ea61:2468db75 > >>>>>>> > >>>>>>> Internal Bitmap : 8 sectors from superblock > >>>>>>> Reshape pos'n : 124311040 (118.55 GiB 127.29 GB) > >>>>>>> Delta Devices : 1 (8->9) > >>>>>>> > >>>>>>> Update Time : Tue Jul 11 23:12:08 2023 > >>>>>>> Bad Block Log : 512 entries available at offset 72 sectors - bad > >>>>>>> blocks present. > >>>>>>> Checksum : 88d8b8fc - correct > >>>>>>> Events : 181105 > >>>>>>> > >>>>>>> Layout : left-symmetric > >>>>>>> Chunk Size : 512K > >>>>>>> > >>>>>>> Device Role : Active device 4 > >>>>>>> Array State : AA.AAAAA. ('A' == active, '.' == missing, 'R' == replacing) > >>>>>>> > >>>>>>> $ sudo mdadm --examine /dev/sdd > >>>>>>> /dev/sdd: > >>>>>>> MBR Magic : aa55 > >>>>>>> Partition[0] : 4294967295 sectors at 1 (type ee) > >>>>>>> $ sudo mdadm --examine /dev/sdd1 > >>>>>>> /dev/sdd1: > >>>>>>> Magic : a92b4efc > >>>>>>> Version : 1.2 > >>>>>>> Feature Map : 0x5 > >>>>>>> Array UUID : 440dc11e:079308b1:131eda79:9a74c670 > >>>>>>> Name : Blyth:0 (local to host Blyth) > >>>>>>> Creation Time : Tue Aug 4 23:47:57 2015 > >>>>>>> Raid Level : raid6 > >>>>>>> Raid Devices : 9 > >>>>>>> > >>>>>>> Avail Dev Size : 5856376832 sectors (2.73 TiB 3.00 TB) > >>>>>>> Array Size : 20497268736 KiB (19.09 TiB 20.99 TB) > >>>>>>> Used Dev Size : 5856362496 sectors (2.73 TiB 3.00 TB) > >>>>>>> Data Offset : 247808 sectors > >>>>>>> Super Offset : 8 sectors > >>>>>>> Unused Space : before=247728 sectors, after=14336 sectors > >>>>>>> State : clean > >>>>>>> Device UUID : a198095b:f54d26a9:deb3be8f:d6de9be1 > >>>>>>> > >>>>>>> Internal Bitmap : 8 sectors from superblock > >>>>>>> Reshape pos'n : 124311040 (118.55 GiB 127.29 GB) > >>>>>>> Delta Devices : 1 (8->9) > >>>>>>> > >>>>>>> Update Time : Tue Jul 11 23:12:08 2023 > >>>>>>> Bad Block Log : 512 entries available at offset 24 sectors > >>>>>>> Checksum : d1471d9d - correct > >>>>>>> Events : 181105 > >>>>>>> > >>>>>>> Layout : left-symmetric > >>>>>>> Chunk Size : 512K > >>>>>>> > >>>>>>> Device Role : Active device 6 > >>>>>>> Array State : AA.AAAAA. ('A' == active, '.' == missing, 'R' == replacing) > >>>>>>> > >>>>>>> $ sudo mdadm --examine /dev/sde > >>>>>>> /dev/sde: > >>>>>>> MBR Magic : aa55 > >>>>>>> Partition[0] : 4294967295 sectors at 1 (type ee) > >>>>>>> $ sudo mdadm --examine /dev/sde1 > >>>>>>> /dev/sde1: > >>>>>>> Magic : a92b4efc > >>>>>>> Version : 1.2 > >>>>>>> Feature Map : 0x5 > >>>>>>> Array UUID : 440dc11e:079308b1:131eda79:9a74c670 > >>>>>>> Name : Blyth:0 (local to host Blyth) > >>>>>>> Creation Time : Tue Aug 4 23:47:57 2015 > >>>>>>> Raid Level : raid6 > >>>>>>> Raid Devices : 9 > >>>>>>> > >>>>>>> Avail Dev Size : 5856376832 sectors (2.73 TiB 3.00 TB) > >>>>>>> Array Size : 20497268736 KiB (19.09 TiB 20.99 TB) > >>>>>>> Used Dev Size : 5856362496 sectors (2.73 TiB 3.00 TB) > >>>>>>> Data Offset : 247808 sectors > >>>>>>> Super Offset : 8 sectors > >>>>>>> Unused Space : before=247720 sectors, after=14336 sectors > >>>>>>> State : clean > >>>>>>> Device UUID : acf7ba2e:35d2fa91:6b12b0ce:33a73af5 > >>>>>>> > >>>>>>> Internal Bitmap : 8 sectors from superblock > >>>>>>> Reshape pos'n : 124311040 (118.55 GiB 127.29 GB) > >>>>>>> Delta Devices : 1 (8->9) > >>>>>>> > >>>>>>> Update Time : Tue Jul 11 23:12:08 2023 > >>>>>>> Bad Block Log : 512 entries available at offset 72 sectors > >>>>>>> Checksum : e05d0278 - correct > >>>>>>> Events : 181105 > >>>>>>> > >>>>>>> Layout : left-symmetric > >>>>>>> Chunk Size : 512K > >>>>>>> > >>>>>>> Device Role : Active device 5 > >>>>>>> Array State : AA.AAAAA. ('A' == active, '.' == missing, 'R' == replacing) > >>>>>>> > >>>>>>> $ sudo mdadm --examine /dev/sdf > >>>>>>> /dev/sdf: > >>>>>>> MBR Magic : aa55 > >>>>>>> Partition[0] : 4294967295 sectors at 1 (type ee) > >>>>>>> $ sudo mdadm --examine /dev/sdf1 > >>>>>>> /dev/sdf1: > >>>>>>> Magic : a92b4efc > >>>>>>> Version : 1.2 > >>>>>>> Feature Map : 0x5 > >>>>>>> Array UUID : 440dc11e:079308b1:131eda79:9a74c670 > >>>>>>> Name : Blyth:0 (local to host Blyth) > >>>>>>> Creation Time : Tue Aug 4 23:47:57 2015 > >>>>>>> Raid Level : raid6 > >>>>>>> Raid Devices : 9 > >>>>>>> > >>>>>>> Avail Dev Size : 5856373760 sectors (2.73 TiB 3.00 TB) > >>>>>>> Array Size : 20497268736 KiB (19.09 TiB 20.99 TB) > >>>>>>> Used Dev Size : 5856362496 sectors (2.73 TiB 3.00 TB) > >>>>>>> Data Offset : 247808 sectors > >>>>>>> Super Offset : 8 sectors > >>>>>>> Unused Space : before=247720 sectors, after=14336 sectors > >>>>>>> State : clean > >>>>>>> Device UUID : 31e7b86d:c274ff45:aa6dab50:2ff058c6 > >>>>>>> > >>>>>>> Internal Bitmap : 8 sectors from superblock > >>>>>>> Reshape pos'n : 124311040 (118.55 GiB 127.29 GB) > >>>>>>> Delta Devices : 1 (8->9) > >>>>>>> > >>>>>>> Update Time : Tue Jul 11 23:12:08 2023 > >>>>>>> Bad Block Log : 512 entries available at offset 72 sectors > >>>>>>> Checksum : 26792cc0 - correct > >>>>>>> Events : 181105 > >>>>>>> > >>>>>>> Layout : left-symmetric > >>>>>>> Chunk Size : 512K > >>>>>>> > >>>>>>> Device Role : Active device 0 > >>>>>>> Array State : AA.AAAAA. ('A' == active, '.' == missing, 'R' == replacing) > >>>>>>> > >>>>>>> $ sudo mdadm --examine /dev/sdg > >>>>>>> /dev/sdg: > >>>>>>> MBR Magic : aa55 > >>>>>>> Partition[0] : 4294967295 sectors at 1 (type ee) > >>>>>>> $ sudo mdadm --examine /dev/sdg1 > >>>>>>> /dev/sdg1: > >>>>>>> Magic : a92b4efc > >>>>>>> Version : 1.2 > >>>>>>> Feature Map : 0x5 > >>>>>>> Array UUID : 440dc11e:079308b1:131eda79:9a74c670 > >>>>>>> Name : Blyth:0 (local to host Blyth) > >>>>>>> Creation Time : Tue Aug 4 23:47:57 2015 > >>>>>>> Raid Level : raid6 > >>>>>>> Raid Devices : 9 > >>>>>>> > >>>>>>> Avail Dev Size : 5856373760 sectors (2.73 TiB 3.00 TB) > >>>>>>> Array Size : 20497268736 KiB (19.09 TiB 20.99 TB) > >>>>>>> Used Dev Size : 5856362496 sectors (2.73 TiB 3.00 TB) > >>>>>>> Data Offset : 247808 sectors > >>>>>>> Super Offset : 8 sectors > >>>>>>> Unused Space : before=247720 sectors, after=14336 sectors > >>>>>>> State : clean > >>>>>>> Device UUID : 74476ce7:4edc23f6:08120711:ba281425 > >>>>>>> > >>>>>>> Internal Bitmap : 8 sectors from superblock > >>>>>>> Reshape pos'n : 124311040 (118.55 GiB 127.29 GB) > >>>>>>> Delta Devices : 1 (8->9) > >>>>>>> > >>>>>>> Update Time : Tue Jul 11 23:12:08 2023 > >>>>>>> Bad Block Log : 512 entries available at offset 72 sectors > >>>>>>> Checksum : 6f67d179 - correct > >>>>>>> Events : 181105 > >>>>>>> > >>>>>>> Layout : left-symmetric > >>>>>>> Chunk Size : 512K > >>>>>>> > >>>>>>> Device Role : Active device 1 > >>>>>>> Array State : AA.AAAAA. ('A' == active, '.' == missing, 'R' == replacing) > >>>>>>> > >>>>>>> $ sudo mdadm --examine /dev/sdh > >>>>>>> /dev/sdh: > >>>>>>> MBR Magic : aa55 > >>>>>>> Partition[0] : 4294967295 sectors at 1 (type ee) > >>>>>>> $ sudo mdadm --examine /dev/sdh1 > >>>>>>> /dev/sdh1: > >>>>>>> Magic : a92b4efc > >>>>>>> Version : 1.2 > >>>>>>> Feature Map : 0xd > >>>>>>> Array UUID : 440dc11e:079308b1:131eda79:9a74c670 > >>>>>>> Name : Blyth:0 (local to host Blyth) > >>>>>>> Creation Time : Tue Aug 4 23:47:57 2015 > >>>>>>> Raid Level : raid6 > >>>>>>> Raid Devices : 9 > >>>>>>> > >>>>>>> Avail Dev Size : 5856373760 sectors (2.73 TiB 3.00 TB) > >>>>>>> Array Size : 20497268736 KiB (19.09 TiB 20.99 TB) > >>>>>>> Used Dev Size : 5856362496 sectors (2.73 TiB 3.00 TB) > >>>>>>> Data Offset : 247808 sectors > >>>>>>> Super Offset : 8 sectors > >>>>>>> Unused Space : before=247720 sectors, after=14336 sectors > >>>>>>> State : clean > >>>>>>> Device UUID : 31c08263:b135f0f5:763bc86b:f81d7296 > >>>>>>> > >>>>>>> Internal Bitmap : 8 sectors from superblock > >>>>>>> Reshape pos'n : 124207104 (118.45 GiB 127.19 GB) > >>>>>>> Delta Devices : 1 (8->9) > >>>>>>> > >>>>>>> Update Time : Tue Jul 11 20:09:14 2023 > >>>>>>> Bad Block Log : 512 entries available at offset 72 sectors - bad > >>>>>>> blocks present. > >>>>>>> Checksum : b7696b68 - correct > >>>>>>> Events : 181089 > >>>>>>> > >>>>>>> Layout : left-symmetric > >>>>>>> Chunk Size : 512K > >>>>>>> > >>>>>>> Device Role : Active device 2 > >>>>>>> Array State : AAAAAAAA. ('A' == active, '.' == missing, 'R' == replacing) > >>>>>>> > >>>>>>> $ sudo mdadm --examine /dev/sdi > >>>>>>> /dev/sdi: > >>>>>>> MBR Magic : aa55 > >>>>>>> Partition[0] : 4294967295 sectors at 1 (type ee) > >>>>>>> $ sudo mdadm --examine /dev/sdi1 > >>>>>>> /dev/sdi1: > >>>>>>> Magic : a92b4efc > >>>>>>> Version : 1.2 > >>>>>>> Feature Map : 0x5 > >>>>>>> Array UUID : 440dc11e:079308b1:131eda79:9a74c670 > >>>>>>> Name : Blyth:0 (local to host Blyth) > >>>>>>> Creation Time : Tue Aug 4 23:47:57 2015 > >>>>>>> Raid Level : raid6 > >>>>>>> Raid Devices : 9 > >>>>>>> > >>>>>>> Avail Dev Size : 5856373760 sectors (2.73 TiB 3.00 TB) > >>>>>>> Array Size : 20497268736 KiB (19.09 TiB 20.99 TB) > >>>>>>> Used Dev Size : 5856362496 sectors (2.73 TiB 3.00 TB) > >>>>>>> Data Offset : 247808 sectors > >>>>>>> Super Offset : 8 sectors > >>>>>>> Unused Space : before=247720 sectors, after=14336 sectors > >>>>>>> State : clean > >>>>>>> Device UUID : ac1063fc:d9d66e6d:f3de33da:b396f483 > >>>>>>> > >>>>>>> Internal Bitmap : 8 sectors from superblock > >>>>>>> Reshape pos'n : 124311040 (118.55 GiB 127.29 GB) > >>>>>>> Delta Devices : 1 (8->9) > >>>>>>> > >>>>>>> Update Time : Tue Jul 11 23:12:08 2023 > >>>>>>> Bad Block Log : 512 entries available at offset 72 sectors > >>>>>>> Checksum : 23b6d024 - correct > >>>>>>> Events : 181105 > >>>>>>> > >>>>>>> Layout : left-symmetric > >>>>>>> Chunk Size : 512K > >>>>>>> > >>>>>>> Device Role : Active device 3 > >>>>>>> Array State : AA.AAAAA. ('A' == active, '.' == missing, 'R' == replacing) > >>>>>>> > >>>>>>> $ sudo mdadm --detail /dev/md0 > >>>>>>> /dev/md0: > >>>>>>> Version : 1.2 > >>>>>>> Raid Level : raid6 > >>>>>>> Total Devices : 9 > >>>>>>> Persistence : Superblock is persistent > >>>>>>> > >>>>>>> State : inactive > >>>>>>> Working Devices : 9 > >>>>>>> > >>>>>>> Delta Devices : 1, (-1->0) > >>>>>>> New Level : raid6 > >>>>>>> New Layout : left-symmetric > >>>>>>> New Chunksize : 512K > >>>>>>> > >>>>>>> Name : Blyth:0 (local to host Blyth) > >>>>>>> UUID : 440dc11e:079308b1:131eda79:9a74c670 > >>>>>>> Events : 181105 > >>>>>>> > >>>>>>> Number Major Minor RaidDevice > >>>>>>> > >>>>>>> - 8 1 - /dev/sda1 > >>>>>>> - 8 129 - /dev/sdi1 > >>>>>>> - 8 113 - /dev/sdh1 > >>>>>>> - 8 97 - /dev/sdg1 > >>>>>>> - 8 81 - /dev/sdf1 > >>>>>>> - 8 65 - /dev/sde1 > >>>>>>> - 8 49 - /dev/sdd1 > >>>>>>> - 8 33 - /dev/sdc1 > >>>>>>> - 8 17 - /dev/sdb1 > >>>>>>> > >>>>>>> $ cat /proc/mdstat > >>>>>>> Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] > >>>>>>> [raid4] [raid10] > >>>>>>> md0 : inactive sdb1[9](S) sdi1[4](S) sdf1[0](S) sdg1[1](S) sdh1[3](S) > >>>>>>> sda1[8](S) sdd1[7](S) sdc1[6](S) sde1[5](S) > >>>>>>> 26353689600 blocks super 1.2 > >>>>>>> > >>>>>>> unused devices: <none> > >>>>>>> > >>>>>>> . > >>>>>>> > >>>>>> > >>>>> > >>>>> . > >>>>> > >>>> > >>> > >>> . > >>> > >> > > > > . > > >