> What Kingston SSD model? === START OF INFORMATION SECTION === Model Family: SandForce Driven SSDs Device Model: KINGSTON SE50S3100G Serial Number: xxxxxxxxxxxxxxxx LU WWN Device Id: xxxxxxxxxxxxxxxx Firmware Version: 611ABBF0 User Capacity: 100,030,242,816 bytes [100 GB] Sector Size: 512 bytes logical/physical Rotation Rate: Solid State Device Form Factor: 2.5 inches Device is: In smartctl database [for details use: -P show] ATA Version is: ATA8-ACS, ACS-2 T13/2015-D revision 3 SATA Version is: SATA 3.0, 6.0 Gb/s (current: 6.0 Gb/s) Local Time is: Tue Aug 4 14:31:36 2020 MSK SMART support is: Available - device has SMART capability. SMART support is: Enabled вт, 4 авг. 2020 г. в 14:17, Eneko Lacunza <elacunza@xxxxxxxxx>: > Hi Vladimir, > > What Kingston SSD model? > > El 4/8/20 a las 12:22, Vladimir Prokofev escribió: > > Here's some more insight into the issue. > > Looks like the load is triggered because of a snaptrim operation. We > have a > > backup pool that serves as Openstack cinder-backup storage, performing > > snapshot backups every night. Old backups are also deleted every night, > so > > snaptrim is initiated. > > This snaptrim increased load on the block.db devices after upgrade, and > > just kills one SSD's performance in particular. It serves as a > block.db/wal > > device for one of the fatter backup pool OSDs which has more PGs placed > > there. > > This is a Kingston SSD, and we see this issue on other Kingston SSD > > journals too, Intel SSD journals are not that affected, though they too > > experience increased load. > > Nevertheless, there're now a lot of read IOPS on block.db devices after > > upgrade that were not there before. > > I wonder how 600 IOPS can destroy SSDs performance that hard. > > > > вт, 4 авг. 2020 г. в 12:54, Vladimir Prokofev <v@xxxxxxxxxxx>: > > > >> Good day, cephers! > >> > >> We've recently upgraded our cluster from 14.2.8 to 14.2.10 release, also > >> performing full system packages upgrade(Ubuntu 18.04 LTS). > >> After that performance significantly dropped, main reason beeing that > >> journal SSDs are now have no merges, huge queues, and increased latency. > >> There's a few screenshots in attachments. This is for an SSD journal > that > >> supports block.db/block.wal for 3 spinning OSDs, and it looks like this > for > >> all our SSD block.db/wal devices across all nodes. > >> Any ideas what may cause that? Maybe I've missed something important in > >> release notes? > >> > > _______________________________________________ > > ceph-users mailing list -- ceph-users@xxxxxxx > > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > > -- > Eneko Lacunza | Tel. 943 569 206 > | Email elacunza@xxxxxxxxx > Director Técnico | Site. https://www.binovo.es > BINOVO IT HUMAN PROJECT S.L | Dir. Astigarragako Bidea, 2 - 2º izda. > Oficina 10-11, 20180 Oiartzun > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx