Hi Mike, > We have similar setup, and I do not test restoring... > How many volumes do you have - one volume on one (*3) disk 10 TB in size > - then 4 volumes? Testing could be quite easy: reset-brick start, then delete&re-create partition/fs/etc., reset-brick commit force - and then watch. We only have 1 big volume over all bricks. Details: Volume Name: shared Type: Distributed-Replicate Number of Bricks: 4 x 3 = 12 Brick1: gluster11:/gluster/bricksda1/shared Brick2: gluster12:/gluster/bricksda1/shared Brick3: gluster13:/gluster/bricksda1/shared Brick4: gluster11:/gluster/bricksdb1/shared Brick5: gluster12:/gluster/bricksdb1/shared Brick6: gluster13:/gluster/bricksdb1/shared Brick7: gluster11:/gluster/bricksdc1/shared Brick8: gluster12:/gluster/bricksdc1/shared Brick9: gluster13:/gluster/bricksdc1/shared Brick10: gluster11:/gluster/bricksdd1/shared Brick11: gluster12:/gluster/bricksdd1_new/shared Brick12: gluster13:/gluster/bricksdd1_new/shared Didn't think about creating more volumes (in order to split data), e.g. 4 volumes with 3*10TB each, or 2 volumes with 6*10TB each. Just curious: after splitting into 2 or more volumes - would that make the volume with the healthy/non-restoring disks better accessable? And only the volume with the once faulty and now restoring disk would be in a "bad mood"? > > Any opinions on that? Maybe it would be better to use more servers and > > smaller disks, but this isn't possible at the moment. > Also interested. We can swap SSDs to HDDs for RAID10, but is it worthless? Yeah, would be interested in how the glusterfs professionsals deal with faulty disks, especially when these are as big as our ones. Thx Hubert _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx https://lists.gluster.org/mailman/listinfo/gluster-users