Re: How do I tell which disk failed?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, 2013-01-08 at 15:38 -0700, Chris Murphy wrote:
> On Jan 8, 2013, at 2:54 PM, Ross Boylan <ross@xxxxxxxxxxxxxxxx> wrote:
> > 
> > I manually specified the current layout of the bigger disks (sdb and c);
> > at least some of the time I specified the exact sector. I picked 34
> > because that seems to be the traditional offset for the first partition
> > (and the one my tool generated when I gave it sizes in grosser units
> > than sectors or told it to start at 0).
> 
> Today 34 is both old and incorrect, so you need to redo the layout.
> 
> > Apparently some disks do a logical to physical remap that includes an
> > offset as well as a change in the sector size.  Should I check for that,
> > or should I just assume that I should start my partitions on sectors
> > that are multiples of 8?
> 
> I know of no disks that change the sector size. It's always 512 logical, 4096 physical for reds. There are supposed to be native 4Kn drives between now and soon, but they aren't switchable between 512e and 4Kn. As for the offset, that still won't work because it'll change the position of your partition map so you'd have to start over anyway, even if it were available, which I don't think it is on a red.
> 
I didn't mean that the disk changed its sector size dynamically, just
that, e.g., it might have physical sectors of 4k but report that it has
(logical) sectors of 512.

I'm not sure what you mean by the offset working.  I'm referring to the
fact that for some drives when you ask for logical sector n you actually
get physical sector n+1, n-2, or something like that.  This implies that
aligning on the logical sectors (meaning the ones the drive reports out)
might misalign on the physical ones.
> So you just need to use a more recent partition tool and repartition the disks correctly.
Correctly = start at multiples of 8?
> 
> 
> 
> 
> 
> > 
> > You also asked what I meant by chatter in the logs about sdb.  Here are
> > some entries from shortly before the system locked up:
> > Jan  6 03:45:24 markov smartd[5368]: Device: /dev/sda, SMART Usage Attribute: 190 Airflow_Temperature_Cel changed from 64 to 65
> > Jan  6 03:45:24 markov smartd[5368]: Device: /dev/sda, SMART Usage Attribute: 194 Temperature_Celsius changed from 36 to 35
> > Jan  6 03:45:25 markov smartd[5368]: Device: /dev/sdb, SMART Usage Attribute: 194 Temperature_Celsius changed from 108 to 109
> 
> smartmontools 5.38 is old, and this red drive isn't in its database, so the data may be interpreted incorrectly. 108 C is very hot. But I wouldn't totally discount it when the drives are all busy on a resync, if you get wildly different Raw_Values for this attribute between sdb and sdc since they're the same drive model.
> 
That report was from before the system crash, when it was probably doing
very little, although disk intensive maintenance such as backups or
indexing the mail spool might have been happening.

I thought 108 was the scaled smart score, which is between 0 and 255
with higher being better.  The raw value of 45 seemed more plausible as
an actual temperature, though I guess there's no guarantee of that.

sdb and sdc have similar numbers for Temperature_Celsius.

On the logs and sign of disk failure, it's quite possible I don't know
what I'm looking for.  Given their size and the fact that the drive
failure seems clear, I think I'll  spare you all the gory details.

Ross


--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux