Re: Many concurrent drive failures - How do I activate pgs?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Mike,

 

I eventually got hold of a customer relations manager at Intel but his attitude was lack luster and Intel never officially responded to any correspondence we sent them. The Intel s4600 drives all passed our standard burn-in tests, they exclusively appear to fail once they handle production BlueStore usage, generally after a couple days use.

 

Intel really didn’t seem interested, even after explaining that the drives were in different physical systems in different data centres and that I had been in contact with another Intel customer who had experienced similar failures in Dell equipment (our servers are pure Intel).

 

 

Perhaps there’s interest in a Lawyer picking up the issue and their attitude. Not advising customers of a known issue which leads to data loss is simply negligent, especially on a product that they tout as being more reliable than spinners and has their Data Centre reliability stamp.

 

I returned the lot and am done with Intel SSDs, will advise as many customers and peers to do the same…

 

 

Regards

David Herselman

 

 

From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf Of Mike Lovell
Sent: Thursday, 22 February 2018 11:19 PM
To: ceph-users@xxxxxxxxxxxxxx
Subject: Re: [ceph-users] Many concurrent drive failures - How do I activate pgs?

 

has anyone tried with the most recent firmwares from intel? i've had a number of s4600 960gb drives that have been waiting for me to get around to adding them to a ceph cluster. this as well as having 2 die almost simultaneously in a different storage box is giving me pause. i noticed that David listed some output showing his ssds were running firmware version SCV10100. the drives i have came with the same one. it looks like SCV10111 is available through the latest isdct package. i'm working through upgrading mine and attempting some burn in testing. just curious if anyone has had any luck there.

 

mike

 

On Thu, Feb 22, 2018 at 9:49 AM, Chris Sarginson <csargiso@xxxxxxxxx> wrote:

Hi Caspar,

 

Sean and I replaced the problematic DC S4600 disks (after all but one had failed) in our cluster with Samsung SM863a disks.  

There was an NDA for new Intel firmware (as mentioned earlier in the thread by David) but given the problems we were experiencing we moved all Intel disks to a single failure domain but were unable to get to deploy additional firmware to test.


The Samsung should fit your requirements.

 

 

Regards

Chris

 

On Thu, 22 Feb 2018 at 12:50 Caspar Smit <casparsmit@xxxxxxxxxxx> wrote:

Hi Sean and David,

 

Do you have any follow ups / news on the Intel DC S4600 case? We are looking into this drives to use as DB/WAL devices for a new to be build cluster.

 

Did Intel provide anything (like new firmware) which should fix the issues you were having or are these drives still unreliable?

 

At the moment we are also looking into the Intel DC S3610 as an alternative which are a step back in performance but should be very reliable.

 

Maybe any other recommendations for a ~200GB 2,5" SATA SSD to use as DB/WAL? (Aiming for ~3 DWPD should be sufficient for DB/WAL?) 

 

Kind regards,

Caspar

 

2018-01-12 15:45 GMT+01:00 Sean Redmond <sean.redmond1@xxxxxxxxx>:

Hi David,

 

To follow up on this I had a 4th drive fail (out of 12) and have opted to order the below disks as a replacement, I have an ongoing case with Intel via the supplier - Will report back anything useful - But I am going to avoid the Intel s4600 2TB SSD's for the moment. 

 

1.92TB Samsung SM863a 2.5" Enterprise SSD, SATA3 6Gb/s, 2-bit MLC V-NAND

 

Regards

Sean Redmond

 

On Wed, Jan 10, 2018 at 11:08 PM, Sean Redmond <sean.redmond1@xxxxxxxxx> wrote:

Hi David,

 

Thanks for your email, they are connected inside Dell R730XD (2.5 inch 24 disk model) in None RAID mode via a perc RAID card.

 

The version of ceph is Jewel with kernel 4.13.X and ubuntu 16.04.

 

Thanks for your feedback on the HGST disks.

 

Thanks

 

On Wed, Jan 10, 2018 at 10:55 PM, David Herselman <dhe@xxxxxxxx> wrote:

Hi Sean,

 

No, Intel’s feedback has been… Pathetic… I have yet to receive anything more than a request to ‘sign’ a non-disclosure agreement, to obtain beta firmware. No official answer as to whether or not one can logically unlock the drives, no answer to my question whether or not Intel publish serial numbers anywhere pertaining to recalled batches and no information pertaining to whether or not firmware updates would address any known issues.

 

This with us being an accredited Intel Gold partner…

 

 

We’ve returned the lot and ended up with 9/12 of the drives failing in the same manner. The replaced drives, which had different serial number ranges, also failed. Very frustrating is that the drives fail in a way that result in unbootable servers, unless one adds ‘rootdelay=240’ to the kernel.

 

 

I would be interested to know what platform your drives were in and whether or not they were connected to a RAID module/card.

 

PS: After much searching we’ve decided to order the NVMe conversion kit and have ordered HGST UltraStar SN200 2.5 inch SFF drives with a 3 DWPD rating.

 

 

Regards

David Herselman

 

From: Sean Redmond [mailto:sean.redmond1@xxxxxxxxx]
Sent: Thursday, 11 January 2018 12:45 AM
To: David Herselman <dhe@xxxxxxxx>
Cc: Christian Balzer <chibi@xxxxxxx>; ceph-users@xxxxxxxxxxxxxx


Subject: Re: [ceph-users] Many concurrent drive failures - How do I activate pgs?

 

Hi,

 

I have a case where 3 out to 12 of these Intel S4600 2TB model failed within a matter of days after being burn-in tested then placed into production.

 

I am interested to know, did you every get any further feedback from the vendor on your issue?

 

Thanks

 

On Thu, Dec 21, 2017 at 1:38 PM, David Herselman <dhe@xxxxxxxx> wrote:

Hi,

I assume this can only be a physical manufacturing flaw or a firmware bug? Do Intel publish advisories on recalled equipment? Should others be concerned about using Intel DC S4600 SSD drives? Could this be an electrical issue on the Hot Swap Backplane or BMC firmware issue? Either way, all pure Intel...

The hole is only 1.3 GB (4 MB x 339 objects) but perfectly striped through images, file systems are subsequently severely damaged.

Is it possible to get Ceph to read in partial data shards? It would provide between 25-75% more yield...


Is there anything wrong with how we've proceeded thus far? Would be nice to reference examples of using ceph-objectstore-tool but documentation is virtually non-existent.

We used another SSD drive to simulate bringing all the SSDs back online. We carved up the drive to provide equal partitions to essentially simulate the original SSDs:
  # Partition a drive to provide 12 x 150GB partitions, eg:
    sdd       8:48   0   1.8T  0 disk
    |-sdd1    8:49   0   140G  0 part
    |-sdd2    8:50   0   140G  0 part
    |-sdd3    8:51   0   140G  0 part
    |-sdd4    8:52   0   140G  0 part
    |-sdd5    8:53   0   140G  0 part
    |-sdd6    8:54   0   140G  0 part
    |-sdd7    8:55   0   140G  0 part
    |-sdd8    8:56   0   140G  0 part
    |-sdd9    8:57   0   140G  0 part
    |-sdd10   8:58   0   140G  0 part
    |-sdd11   8:59   0   140G  0 part
    +-sdd12   8:60   0   140G  0 part


  Pre-requisites:
    ceph osd set noout;
    apt-get install uuid-runtime;


  for ID in `seq 24 35`; do
    UUID=`uuidgen`;
    OSD_SECRET=`ceph-authtool --gen-print-key`;
    DEVICE='/dev/sdd'$[$ID-23]; # 24-23 = /dev/sdd1, 35-23 = /dev/sdd12
    echo "{\"cephx_secret\": \"$OSD_SECRET\"}" | ceph osd new $UUID $ID -i - -n client.bootstrap-osd -k /var/lib/ceph/bootstrap-osd/ceph.keyring;
    mkdir /var/lib/ceph/osd/ceph-$ID;
    mkfs.xfs $DEVICE;
    mount $DEVICE /var/lib/ceph/osd/ceph-$ID;
    ceph-authtool --create-keyring /var/lib/ceph/osd/ceph-$ID/keyring --name osd.$ID --add-key $OSD_SECRET;
    ceph-osd -i $ID --mkfs --osd-uuid $UUID;
    chown -R ceph:ceph /var/lib/ceph/osd/ceph-$ID;
    systemctl enable ceph-osd@$ID;
    systemctl start ceph-osd@$ID;
  done


Once up we imported previous exports of empty head files in to 'real' OSDs:
  kvm5b:
    systemctl stop ceph-osd@8;
    ceph-objectstore-tool --op import --pgid 7.4s0 --data-path /var/lib/ceph/osd/ceph-8 --journal-path /var/lib/ceph/osd/ceph-8/journal --file /var/lib/vz/template/ssd_recovery/osd8_7.4s0.export;
    chown ceph:ceph -R /var/lib/ceph/osd/ceph-8;
    systemctl start ceph-osd@8;
  kvm5f:
    systemctl stop ceph-osd@23;
    ceph-objectstore-tool --op import --pgid 7.fs0 --data-path /var/lib/ceph/osd/ceph-23 --journal-path /var/lib/ceph/osd/ceph-23/journal --file /var/lib/vz/template/ssd_recovery/osd23_7.fs0.export;
    chown ceph:ceph -R /var/lib/ceph/osd/ceph-23;
    systemctl start ceph-osd@23;


Bulk import previously exported objects:
    cd /var/lib/vz/template/ssd_recovery;
    for FILE in `ls -1A osd*_*.export | grep -Pv '^osd(8|23)_'`; do
      OSD=`echo $FILE | perl -pe 's/^osd(\d+).*/\1/'`;
      PGID=`echo $FILE | perl -pe 's/^osd\d+_(.*?).export/\1/g'`;
      echo -e "systemctl stop ceph-osd@$OSD\t ceph-objectstore-tool --op import --pgid $PGID --data-path /var/lib/ceph/osd/ceph-$OSD --journal-path /var/lib/ceph/osd/ceph-$OSD/journal --file /var/lib/vz/template/ssd_recovery/osd"$OSD"_$PGID.export";
    done | sort

Sample output (this will wrap):
systemctl stop ceph-osd@27       ceph-objectstore-tool --op import --pgid 7.4s3 --data-path /var/lib/ceph/osd/ceph-27 --journal-path /var/lib/ceph/osd/ceph-27/journal --file /var/lib/vz/template/ssd_recovery/osd27_7.4s3.export
systemctl stop ceph-osd@27       ceph-objectstore-tool --op import --pgid 7.fs5 --data-path /var/lib/ceph/osd/ceph-27 --journal-path /var/lib/ceph/osd/ceph-27/journal --file /var/lib/vz/template/ssd_recovery/osd27_7.fs5.export
systemctl stop ceph-osd@30       ceph-objectstore-tool --op import --pgid 7.fs4 --data-path /var/lib/ceph/osd/ceph-30 --journal-path /var/lib/ceph/osd/ceph-30/journal --file /var/lib/vz/template/ssd_recovery/osd30_7.fs4.export
systemctl stop ceph-osd@31       ceph-objectstore-tool --op import --pgid 7.4s2 --data-path /var/lib/ceph/osd/ceph-31 --journal-path /var/lib/ceph/osd/ceph-31/journal --file /var/lib/vz/template/ssd_recovery/osd31_7.4s2.export
systemctl stop ceph-osd@32       ceph-objectstore-tool --op import --pgid 7.4s4 --data-path /var/lib/ceph/osd/ceph-32 --journal-path /var/lib/ceph/osd/ceph-32/journal --file /var/lib/vz/template/ssd_recovery/osd32_7.4s4.export
systemctl stop ceph-osd@32       ceph-objectstore-tool --op import --pgid 7.fs2 --data-path /var/lib/ceph/osd/ceph-32 --journal-path /var/lib/ceph/osd/ceph-32/journal --file /var/lib/vz/template/ssd_recovery/osd32_7.fs2.export
systemctl stop ceph-osd@34       ceph-objectstore-tool --op import --pgid 7.4s5 --data-path /var/lib/ceph/osd/ceph-34 --journal-path /var/lib/ceph/osd/ceph-34/journal --file /var/lib/vz/template/ssd_recovery/osd34_7.4s5.export
systemctl stop ceph-osd@34       ceph-objectstore-tool --op import --pgid 7.fs1 --data-path /var/lib/ceph/osd/ceph-34 --journal-path /var/lib/ceph/osd/ceph-34/journal --file /var/lib/vz/template/ssd_recovery/osd34_7.fs1.export


Reset permissions and then started the OSDs:
for OSD in 27 30 31 32 34; do
  chown -R ceph:ceph /var/lib/ceph/osd/ceph-$OSD;
  systemctl start ceph-osd@$OSD;
done


Then finally started all the OSDs... Now to hope that Intel have a way of accessing drives that are in a 'disable logical state'.



The imports succeed, herewith a link to the output after running an import for placement group 7.4s2 on OSD 31:
  https://drive.google.com/open?id=1-Jo1jmrWrGLO2OgflacGPlEf2p32Y4hn

Sample snippet:
    Write 1#7:fffcd2ec:::rbd_data.4.be8e9974b0dc51.0000000000002869:head#
    snapset 0=[]:{}
    Write 1#7:fffd4823:::rbd_data.4.ba24ef2ae8944a.000000000000a2b0:head#
    snapset 0=[]:{}
    Write 1#7:fffd6fb6:::benchmark_data_kvm5b_20945_object14722:head#
    snapset 0=[]:{}
    Write 1#7:ffffa069:::rbd_data.4.ba24ef2ae8944a.000000000000aea9:head#
    snapset 0=[]:{}
    Import successful


Data does get written, I can tell by the size of the FileStore mount points:
  [root@kvm5b ssd_recovery]# df -h | grep -P 'ceph-(27|30|31|32|34)$'
  /dev/sdd4       140G  5.2G  135G   4% /var/lib/ceph/osd/ceph-27
  /dev/sdd7       140G   14G  127G  10% /var/lib/ceph/osd/ceph-30
  /dev/sdd8       140G   14G  127G  10% /var/lib/ceph/osd/ceph-31
  /dev/sdd9       140G   22G  119G  16% /var/lib/ceph/osd/ceph-32
  /dev/sdd11      140G   22G  119G  16% /var/lib/ceph/osd/ceph-34


How do I tell Ceph to read these object shards?



PS: It's probably a good idea to reweight the OSDs to 0 before starting again. This should prevent data flowing on to them, if they are not in a different device class or other crush selection ruleset. Ie:
  for OSD in `seq 24 35`; do
    ceph osd crush reweight osd.$OSD 0;
  done


Regards
David Herselman

-----Original Message-----

From: David Herselman
Sent: Thursday, 21 December 2017 3:49 AM
To: 'Christian Balzer' <chibi@xxxxxxx>; ceph-users@xxxxxxxxxxxxxx
Subject: RE: [ceph-users] Many concurrent drive failures - How do I activate pgs?

Hi Christian,

Thanks for taking the time, I haven't been contacted by anyone yet but managed to get the down placement groups cleared by exporting 7.4s0 and 7.fs0 and then marking them as complete on the surviving OSDs:
    kvm5c:
      ceph-objectstore-tool --op export --pgid 7.4s0 --data-path /var/lib/ceph/osd/ceph-8 --journal-path /var/lib/ceph/osd/ceph-8/journal --file /var/lib/vz/template/ssd_recovery/osd8_7.4s0.export;
      ceph-objectstore-tool --op mark-complete --data-path /var/lib/ceph/osd/ceph-8 --journal-path /var/lib/ceph/osd/ceph-8/journal --pgid 7.4s0;
    kvm5f:
      ceph-objectstore-tool --op export --pgid 7.fs0 --data-path /var/lib/ceph/osd/ceph-23 --journal-path /var/lib/ceph/osd/ceph-23/journal --file /var/lib/vz/template/ssd_recovery/osd23_7.fs0.export;
      ceph-objectstore-tool --op mark-complete --data-path /var/lib/ceph/osd/ceph-23 --journal-path /var/lib/ceph/osd/ceph-23/journal --pgid 7.fs0;

This would presumably simply punch holes in the RBD images but at least we can copy them out of that pool and hope that Intel can somehow unlock the drives for us to then export/import objects.


To answer your questions though, we have 6 near identical Intel Wildcat Pass 1U servers and have Proxmox loaded on them. Proxmox uses a Debian 9 base with the Ubuntu kernel, for which they apply cherry picked kernel patches (eg Intel NIC driver updates, vhost perf regression and mem-leak fixes, etc):

kvm5a:
       Intel R1208WTTGSR System (serial: BQWS55091014)
       Intel S2600WTTR Motherboard (serial: BQWL54950385, BIOS ID: SE5C610.86B.01.01.0021.032120170601)
       2 x Intel Xeon E5-2640v4 2.4GHz (HT disabled)
       24 x Micron 8GB DDR4 2133MHz (24 x 18ASF1G72PZ-2G1B1)
       Intel AXX10GBNIA I/O Module
kvm5b:
       Intel R1208WTTGS System (serial: BQWS53890178)
       Intel S2600WTT Motherboard (serial: BQWL52550359, BIOS ID: SE5C610.86B.01.01.0021.032120170601)
       2 x Intel Xeon E5-2640v4 2.4GHz (HT enabled)
       4 x Micron 64GB DDR4 2400MHz LR-DIMM (4 x 72ASS8G72LZ-2G3B2)
       Intel AXX10GBNIA I/O Module
kvm5c:
       Intel R1208WT2GS System (serial: BQWS50490279)
       Intel S2600WT2 Motherboard (serial: BQWL44650203, BIOS ID: SE5C610.86B.01.01.0021.032120170601)
       2 x Intel Xeon E5-2640v3 2.6GHz (HT enabled)
       4 x Micron 64GB DDR4 2400MHz LR-DIMM (4 x 72ASS8G72LZ-2G3B2)
       Intel AXX10GBNIA I/O Module
kvm5d:
       Intel R1208WTTGSR System (serial: BQWS62291318)
       Intel S2600WTTR Motherboard (serial: BQWL61855187, BIOS ID: SE5C610.86B.01.01.0021.032120170601)
       2 x Intel Xeon E5-2640v4 2.4GHz (HT enabled)
       4 x Micron 64GB DDR4 2400MHz LR-DIMM (4 x 72ASS8G72LZ-2G3B2)
       Intel AXX10GBNIA I/O Module
kvm5e:
       Intel R1208WTTGSR System (serial: BQWS64290162)
       Intel S2600WTTR Motherboard (serial: BQWL63953066, BIOS ID: SE5C610.86B.01.01.0021.032120170601)
       2 x Intel Xeon E5-2640v4 2.4GHz (HT enabled)
       4 x Micron 64GB DDR4 2400MHz LR-DIMM (4 x 72ASS8G72LZ-2G3B2)
       Intel AXX10GBNIA I/O Module
kvm5f:
       Intel R1208WTTGSR System (serial: BQWS71790632)
       Intel S2600WTTR Motherboard (serial: BQWL71050622, BIOS ID: SE5C610.86B.01.01.0021.032120170601)
       2 x Intel Xeon E5-2640v4 2.4GHz (HT enabled)
       4 x Micron 64GB DDR4 2400MHz LR-DIMM (4 x 72ASS8G72LZ-2G3B2)
       Intel AXX10GBNIA I/O Module

Summary:
  * 5b has an Intel S2600WTT, 5c has an Intel S2600WT2, all others have S2600WTTR Motherboards
  * 5a has ECC Registered Dual Rank DDR DIMMs, all others have ECC LoadReduced-DIMMs
  * 5c has an Intel X540-AT2 10 GbE adapter as the on-board NICs are only 1 GbE


Each system has identical discs:
  * 2 x 480 GB Intel SSD DC S3610 (SSDSC2BX480G4) - partitioned as software RAID1 OS volume and Ceph FileStore journals (spinners)
  * 4 x 2 TB Seagate discs (ST2000NX0243) - Ceph FileStore OSDs (journals in S3610 partitions)
  * 2 x 1.9 TB Intel SSD DC S4600 (SSDSC2KG019T7) - Ceph BlueStore OSDs (problematic)


Additional information:
  * All drives are directly attached to the on-board AHCI SATA controllers, via the standard 2.5 inch drive chassis hot-swap bays.
  * We added 12 x 1.9 TB SSD DC S4600 drives last week Thursday, 2 in each system's slots 7 & 8
  * Systems have been operating with existing Intel SSD DC S3610 and 2 TB Seagate discs for over a year; we added the most recent node (kvm5f) on the 23rd of November.
  * 6 of the 12 Intel SSD DC S4600 drives failed in less than 100 hours.
  * They work perfectly until they suddenly stop responding and are thereafter, even with us physically shutting down the server and powering it back up again, completely inaccessible. Intel diagnostic tool reports 'logically locked'.


Drive failures appear random to me:
    kvm5a - bay 7 - offline - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, SerialNo=BTYM739208851P9DGN
    kvm5a - bay 8 - online  - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, SerialNo=PHYM727602TM1P9DGN
    kvm5b - bay 7 - offline - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, SerialNo=PHYM7276031E1P9DGN
    kvm5b - bay 8 - online  - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, SerialNo=BTYM7392087W1P9DGN
    kvm5c - bay 7 - offline - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, SerialNo=BTYM739200ZJ1P9DGN
    kvm5c - bay 8 - offline - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, SerialNo=BTYM7392088B1P9DGN
    kvm5d - bay 7 - offline - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, SerialNo=BTYM738604Y11P9DGN
    kvm5d - bay 8 - online  - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, SerialNo=PHYM727603181P9DGN
    kvm5e - bay 7 - online  - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, SerialNo=BTYM7392013B1P9DGN
    kvm5e - bay 8 - offline - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, SerialNo=BTYM7392087E1P9DGN
    kvm5f - bay 7 - online  - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, SerialNo=BTYM739208721P9DGN
    kvm5f - bay 8 - online  - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, SerialNo=BTYM739208C41P9DGN


Intel SSD Data Center Tool reports:
C:\isdct>isdct.exe show -intelssd

- Intel SSD DC S4600 Series PHYM7276031E1P9DGN -

Bootloader : Property not found
DevicePath : \\\\.\\PHYSICALDRIVE1
DeviceStatus : Selected drive is in a disable logical state.
Firmware : SCV10100
FirmwareUpdateAvailable : Please contact Intel Customer Support for further assistance at the following website: http://www.intel.com/go/ssdsupport.
Index : 0
ModelNumber : INTEL SSDSC2KG019T7
ProductFamily : Intel SSD DC S4600 Series SerialNumber : PHYM7276031E1P9DGN


C:\isdct>isdct show -a -intelssd 0

- Intel SSD DC S4600 Series PHYM7276031E1P9DGN -

AccessibleMaxAddressSupported : True
AggregationThreshold : Selected drive is in a disable logical state.
AggregationTime : Selected drive is in a disable logical state.
ArbitrationBurst : Selected drive is in a disable logical state.
BusType : 11
CoalescingDisable : Selected drive is in a disable logical state.
ControllerCompatibleIDs : PCI\\VEN_8086&DEV_8C02&REV_05PCI\\VEN_8086&DEV_8C02PCI\\VEN_8086&CC_010601PCI\\VEN_8086&CC_0106PCI\\VEN_8086PCI\\CC_010601PCI\\CC_0106
ControllerDescription : @mshdc.inf,%pci\\cc_010601.devicedesc%;Standard SATA AHCI Controller ControllerID : PCI\\VEN_8086&DEV_8C02&SUBSYS_78461462&REV_05\\3&11583659&0&FA
ControllerIDEMode : False
ControllerManufacturer : @mshdc.inf,%ms-ahci%;Standard SATA AHCI Controller ControllerService : storahci DIPMEnabled : False DIPMSupported : False DevicePath : \\\\.\\PHYSICALDRIVE1 DeviceStatus : Selected drive is in a disable logical state.
DigitalFenceSupported : False
DownloadMicrocodePossible : True
DriverDescription : Standard SATA AHCI Controller DriverMajorVersion : 10 DriverManufacturer : Standard SATA AHCI Controller DriverMinorVersion : 0 DriverVersion : 10.0.16299.98 DynamicMMIOEnabled : The selected drive does not support this feature.
EnduranceAnalyzer : Selected drive is in a disable logical state.
ErrorString : *BAD_CONTEXT_2020 F4
Firmware : SCV10100
FirmwareUpdateAvailable : Please contact Intel Customer Support for further assistance at the following website: http://www.intel.com/go/ssdsupport.
HDD : False
HighPriorityWeightArbitration : Selected drive is in a disable logical state.
IEEE1667Supported : False
IOCompletionQueuesRequested : Selected drive is in a disable logical state.
IOSubmissionQueuesRequested : Selected drive is in a disable logical state.
Index : 0
Intel : True
IntelGen3SATA : True
IntelNVMe : False
InterruptVector : Selected drive is in a disable logical state.
IsDualPort : False
LatencyTrackingEnabled : Selected drive is in a disable logical state.
LowPriorityWeightArbitration : Selected drive is in a disable logical state.
Lun : 0
MaximumLBA : 3750748847
MediumPriorityWeightArbitration : Selected drive is in a disable logical state.
ModelNumber : INTEL SSDSC2KG019T7
NVMePowerState : Selected drive is in a disable logical state.
NativeMaxLBA : Selected drive is in a disable logical state.
OEM : Generic
OpalState : Selected drive is in a disable logical state.
PLITestTimeInterval : Selected drive is in a disable logical state.
PNPString : SCSI\\DISK&VEN_INTEL&PROD_SSDSC2KG019T7\\4&2BE6C224&0&010000
PathID : 1
PhySpeed : Selected drive is in a disable logical state.
PhysicalSectorSize : Selected drive is in a disable logical state.
PhysicalSize : 1920383410176
PowerGovernorAveragePower : Selected drive is in a disable logical state.
PowerGovernorBurstPower : Selected drive is in a disable logical state.
PowerGovernorMode : Selected drive is in a disable logical state.
Product : Youngsville
ProductFamily : Intel SSD DC S4600 Series ProductProtocol : ATA ReadErrorRecoveryTimer : Selected drive is in a disable logical state.
RemoteSecureEraseSupported : False
SCSIPortNumber : 0
SMARTEnabled : True
SMARTHealthCriticalWarningsConfiguration : Selected drive is in a disable logical state.
SMARTSelfTestSupported : True
SMBusAddress : Selected drive is in a disable logical state.
SSCEnabled : False
SanitizeBlockEraseSupported : False
SanitizeCryptoScrambleSupported : True
SanitizeSupported : True
SataGen1 : True
SataGen2 : True
SataGen3 : True
SataNegotiatedSpeed : Unknown
SectorSize : 512
SecurityEnabled : False
SecurityFrozen : False
SecurityLocked : False
SecuritySupported : False
SerialNumber : PHYM7276031E1P9DGN
TCGSupported : False
TargetID : 0
TempThreshold : Selected drive is in a disable logical state.
TemperatureLoggingInterval : Selected drive is in a disable logical state.
TimeLimitedErrorRecovery : Selected drive is in a disable logical state.
TrimSize : 4
TrimSupported : True
VolatileWriteCacheEnabled : Selected drive is in a disable logical state.
WWID : 3959312879584368077
WriteAtomicityDisableNormal : Selected drive is in a disable logical state.
WriteCacheEnabled : True
WriteCacheReorderingStateEnabled : Selected drive is in a disable logical state.
WriteCacheState : Selected drive is in a disable logical state.
WriteCacheSupported : True
WriteErrorRecoveryTimer : Selected drive is in a disable logical state.



SMART information is inaccessible, overall status is failed. Herewith the stats from a partner disc which was still working when the others failed:
Device Model:     INTEL SSDSC2KG019T7
Serial Number:    PHYM727602TM1P9DGN
LU WWN Device Id: 5 5cd2e4 14e1636bb
Firmware Version: SCV10100
User Capacity:    1,920,383,410,176 bytes [1.92 TB]
Sector Sizes:     512 bytes logical, 4096 bytes physical
Rotation Rate:    Solid State Device
Form Factor:      2.5 inches
Device is:        Not in smartctl database [for details use: -P showall]
ATA Version is:   ACS-3 T13/2161-D revision 5
SATA Version is:  SATA 3.2, 6.0 Gb/s (current: 6.0 Gb/s)
Local Time is:    Mon Dec 18 19:33:51 2017 SAST
SMART support is: Available - device has SMART capability.
SMART support is: Enabled

ID# ATTRIBUTE_NAME          FLAG     VALUE WORST THRESH TYPE      UPDATED  WHEN_FAILED RAW_VALUE
  5 Reallocated_Sector_Ct   0x0032   100   100   000    Old_age   Always       -       0
  9 Power_On_Hours          0x0032   100   100   000    Old_age   Always       -       98
12 Power_Cycle_Count       0x0032   100   100   000    Old_age   Always       -       3
170 Unknown_Attribute       0x0033   100   100   010    Pre-fail  Always       -       0
171 Unknown_Attribute       0x0032   100   100   000    Old_age   Always       -       1
172 Unknown_Attribute       0x0032   100   100   000    Old_age   Always       -       0
174 Unknown_Attribute       0x0032   100   100   000    Old_age   Always       -       0
175 Program_Fail_Count_Chip 0x0033   100   100   010    Pre-fail  Always       -       17567121432
183 Runtime_Bad_Block       0x0032   100   100   000    Old_age   Always       -       0
184 End-to-End_Error        0x0033   100   100   090    Pre-fail  Always       -       0
187 Reported_Uncorrect      0x0032   100   100   000    Old_age   Always       -       0
190 Airflow_Temperature_Cel 0x0022   077   076   000    Old_age   Always       -       23 (Min/Max 17/29)
192 Power-Off_Retract_Count 0x0032   100   100   000    Old_age   Always       -       0
194 Temperature_Celsius     0x0022   100   100   000    Old_age   Always       -       23
197 Current_Pending_Sector  0x0012   100   100   000    Old_age   Always       -       0
199 UDMA_CRC_Error_Count    0x003e   100   100   000    Old_age   Always       -       0
225 Unknown_SSD_Attribute   0x0032   100   100   000    Old_age   Always       -       14195
226 Unknown_SSD_Attribute   0x0032   100   100   000    Old_age   Always       -       0
227 Unknown_SSD_Attribute   0x0032   100   100   000    Old_age   Always       -       42
228 Power-off_Retract_Count 0x0032   100   100   000    Old_age   Always       -       5905
232 Available_Reservd_Space 0x0033   100   100   010    Pre-fail  Always       -       0
233 Media_Wearout_Indicator 0x0032   100   100   000    Old_age   Always       -       0
234 Unknown_Attribute       0x0032   100   100   000    Old_age   Always       -       0
241 Total_LBAs_Written      0x0032   100   100   000    Old_age   Always       -       14195
242 Total_LBAs_Read         0x0032   100   100   000    Old_age   Always       -       10422
243 Unknown_Attribute       0x0032   100   100   000    Old_age   Always       -       41906


Media wear out : 0% used
LBAs written: 14195
Power on hours: <100
Power cycle count: once at the factory, once at our offices to check if there was newer firmware (there wasn't) and once when we restarted the node to see if it could then access a failed drive.


Regards
David Herselman


-----Original Message-----
From: Christian Balzer [mailto:chibi@xxxxxxx]
Sent: Thursday, 21 December 2017 3:24 AM
To: ceph-users@xxxxxxxxxxxxxx
Cc: David Herselman <dhe@xxxxxxxx>
Subject: Re: [ceph-users] Many concurrent drive failures - How do I activate pgs?

Hello,

first off, I don't have anything to add to your conclusions of the current status, alas there are at least 2 folks here on the ML making a living from Ceph disaster recovery, so I hope you have been contacted already.

Now once your data is safe or you have a moment, I and others here would probably be quite interested in some more details, see inline below.

On Wed, 20 Dec 2017 22:25:23 +0000 David Herselman wrote:

[snip]
>
> We've happily been running a 6 node cluster with 4 x FileStore HDDs per node (journals on SSD partitions) for over a year and recently upgraded all nodes to Debian 9, Ceph Luminous 12.2.2 and kernel 4.13.8. We ordered 12 x Intel DC S4600 SSDs which arrived last week so we added two per node on Thursday evening and brought them up as BlueStore OSDs. We had proactively updated our existing pools to reference only devices classed as 'hdd', so that we could move select images over to ssd replicated and erasure coded pools.
>
Could you tell us more about that cluster, as in HW, how are the SSDs connected and FW version of the controller if applicable.

Kernel 4.13.8 suggests that this is a handrolled, upstream kernel.
While not necessarily related I'll note that as far as Debian kernels (which are very lightly if at all patched) are concerned, nothing beyond
4.9 has been working to my satisfaction.
4.11 still worked, but 4.12 crash-reboot-looped on all my Supermicro X10 machines (quite a varied selection).
The current 4.13.13 backport boots on some of those machines, but still throws errors with the EDAC devices, which works fine with 4.9.

4.14 is known to happily destroy data if used with bcache and even if one doesn't use that it should give you pause.

> We were pretty diligent and downloaded Intel's Firmware Update Tool and validated that each new drive had the latest available firmware before installing them in the nodes. We did numerous benchmarks on Friday and eventually moved some images over to the new storage pools. Everything was working perfectly and extensive tests on Sunday showed excellent performance. Sunday night one of the new SSDs died and Ceph replicated and redistributed data accordingly, then another failed in the early hours of Monday morning and Ceph did what it needed to.
>
> We had the two failed drives replaced by 11am and Ceph was up to 2/4918587 objects degraded (0.000%) when a third drive failed. At this point we updated the crush maps for the rbd_ssd and ec_ssd pools and set the device class to 'hdd', to essentially evacuate everything off the SSDs. Other SSDs then failed at 3:22pm, 4:19pm, 5:49pm and 5:50pm. We've ultimately lost half the Intel S4600 drives, which are all completely inaccessible. Our status at 11:42pm Monday night was: 1/1398478 objects unfound (0.000%) and 339/4633062 objects degraded (0.007%).
>
The relevant logs when and how those SSDs failed would be interesting.
Was the distribution of the failed SSDs random among the cluster?
Are you running smartd and did it have something to say?

Completely inaccessible sounds a lot like the infamous "self-bricking" of Intel SSDs when they discover something isn't right, or they don't like the color scheme of the server inside (^.^).

I'm using quite a lot of Intel SSDs and had only one "fatal" incident.
A DC S3700 detected that its powercap had failed, but of course kept working fine. Until a reboot was need, when it promptly bricked itself, data inaccessible, SMART reporting barely that something was there.

So one wonders what caused your SSDs to get their knickers in such a twist.
Are the survivors showing any unusual signs in their SMART output?

Of course what your vendor/Intel will have to say will also be of interest. ^o^

Regards,

Christian
--
Christian Balzer        Network/Systems Engineer
chibi@xxxxxxx           Rakuten Communications
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

 

 

 


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

 

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

 

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux