Re: Using XFS and LVM backends together on the same cluster and hosts

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello Anthony. Good to hear from you.

You miss understood somehow, I don't use filestore, I have XFS based
bluestore and LVM based bluestore mixed in the same Nautilus cluster.

Yes NVME's are mdadm raid1. I'm %100 agree with you and when I'm gonna
rebuild, I will pin 1/2 to Nvme1 and 2/2 to Nvme2 without mirror.
I build this system with raid1 nvme because it was one time project and I
wanted to be safe rather than fast.




Anthony D'Atri <anthony.datri@xxxxxxxxx>, 30 Eyl 2024 Pzt, 22:10 tarihinde
şunu yazdı:

> BlueStore vs Filestore doesn’t matter beyond each OSD.  Filestore is very
> deprecated so you’ll want to redeploy any Filestore OSDs when you can.
>  `ceph osd metadata` can survey.
>
> I’ve had multiple issues over time with the MG spinners fwiw.   For what
> SAS spinners cost with some effort you can get SATA SSDs for legacy
> systems.   Do you have those NVMe drives for WAL+DB mirrored?   You
> generally don’t want to go higher than 10:1
>
> > On Sep 30, 2024, at 3:00 PM, Özkan Göksu <ozkangksu@xxxxxxxxx> wrote:
> >
> > Hello folks! I hope you are doing well :)
> >
> > I have a general question about XFS and LVM backend OSD performance
> > and possible effects if they are used together in the same pool.
> >
> > I built a cluster 5 years ago with Nautilus and I used the XFS backend
> for
> > OSD's.
> > After 5 years they reached me back with ERR state and I see 10K++ slow
> ops,
> > 29 incomplete PG, 2 inconsistent PG, 34300 unfound objects and 1 osd down
> > due to compactions problem.
> >
> > At the first check-up I found some drives were replaced by others and
> they
> > used LVM backend for the replaced drives without using wal+db on nvme.
> > In the cluster I have mostly XFS backend drives wal+db on nvme and some
> LVM
> > drives without wal+db.
> >
> > We have Cephfs and RBD pools on SSD drives and 8+2 EC pool for RGW S3
> > workload. RGW index stored on SSD pool.
> > 10 nodes: with
> > - 21 x 16TB Toshiba MG08CA16TEY, Firmware:EJ09 | 8+2 EC RGW DATA POOL
> > - 3 x 960GB MZILS960HEHP/007 Firm: GXL0 | Rep 2 RGW index pool
> > - 2 x PM1725B 1.6T PCI-E NVME | 50G WAL+DB for 21x HDD
> >
> > Total HDD raw size 2.8PiB, SSD size 26TiB
> >
> > I started fixing all the problems one by one and I'm gonna recreate these
> > LVM drives without wal+db and I wonder 2 questions:
> > 1- Are there any speed or latency differences on XFS and LVM backend
> OSD's
> > for 16TB 7200rpm NL-SAS drives.
> > 2- Mixing XFS and LVM backend on the same cluster does have any
> > negative effect or problems?
> >
> > Best regards:
> >
> > Extra note: If you wonder please check the LSBLK output for 1/10 server:
> > NODE-01# lsblk
> > NAME
> >                           MAJ:MIN RM   SIZE RO TYPE  MOUNTPOINT
> > sda
> >                            8:0    0  14.6T  0 disk
> > ├─sda1
> >                             8:1    0   100M  0 part
> > /var/lib/ceph/osd/ceph-180
> > └─sda2
> >                             8:2    0  14.6T  0 part
> > sdb
> >                            8:16   0  14.6T  0 disk
> > ├─sdb1
> >                             8:17   0   100M  0 part
> > /var/lib/ceph/osd/ceph-181
> > └─sdb2
> >                             8:18   0  14.6T  0 part
> > sdc
> >                            8:32   0  14.6T  0 disk
> > ├─sdc1
> >                             8:33   0   100M  0 part
> > /var/lib/ceph/osd/ceph-182
> > └─sdc2
> >                             8:34   0  14.6T  0 part
> > sdd
> >                            8:48   0  14.6T  0 disk
> > ├─sdd1
> >                             8:49   0   100M  0 part
> > /var/lib/ceph/osd/ceph-183
> > └─sdd2
> >                             8:50   0  14.6T  0 part
> > sde
> >                            8:64   0  14.6T  0 disk
> > ├─sde1
> >                             8:65   0   100M  0 part
> > /var/lib/ceph/osd/ceph-185
> > └─sde2
> >                             8:66   0  14.6T  0 part
> > sdf
> >                            8:80   0  14.6T  0 disk
> >
> └─ceph--ef5bd394--8dc9--46a8--a244--0c5d3c1400e3-osd--block--b69c0802--9634--43a5--b4a9--0f36cd8690c5
> > 253:2    0  14.6T  0 lvm
> > sdg
> >                            8:96   0  14.6T  0 disk
> > ├─sdg1
> >                             8:97   0   100M  0 part
> > /var/lib/ceph/osd/ceph-186
> > └─sdg2
> >                             8:98   0  14.6T  0 part
> > sdh
> >                            8:112  0  14.6T  0 disk
> > ├─sdh1
> >                             8:113  0   100M  0 part
> > /var/lib/ceph/osd/ceph-187
> > └─sdh2
> >                             8:114  0  14.6T  0 part
> > sdi
> >                            8:128  0  14.6T  0 disk
> > ├─sdi1
> >                             8:129  0   100M  0 part
> > /var/lib/ceph/osd/ceph-188
> > └─sdi2
> >                             8:130  0  14.6T  0 part
> > sdj
> >                            8:144  0  14.6T  0 disk
> > ├─sdj1
> >                             8:145  0   100M  0 part
> > /var/lib/ceph/osd/ceph-189
> > └─sdj2
> >                             8:146  0  14.6T  0 part
> > sdk
> >                            8:160  0  14.6T  0 disk
> > ├─sdk1
> >                             8:161  0   100M  0 part
> > /var/lib/ceph/osd/ceph-190
> > └─sdk2
> >                             8:162  0  14.6T  0 part
> > sdl
> >                            8:176  0  14.6T  0 disk
> > ├─sdl1
> >                             8:177  0   100M  0 part
> > /var/lib/ceph/osd/ceph-191
> > └─sdl2
> >                             8:178  0  14.6T  0 part
> > sdm
> >                            8:192  0  14.6T  0 disk
> > ├─sdm1
> >                             8:193  0   100M  0 part
> > /var/lib/ceph/osd/ceph-192
> > └─sdm2
> >                             8:194  0  14.6T  0 part
> > sdn
> >                            8:208  0  14.6T  0 disk
> > ├─sdn1
> >                             8:209  0   100M  0 part
> > /var/lib/ceph/osd/ceph-193
> > └─sdn2
> >                             8:210  0  14.6T  0 part
> > sdo
> >                            8:224  0  14.6T  0 disk
> > ├─sdo1
> >                             8:225  0   100M  0 part
> > /var/lib/ceph/osd/ceph-194
> > └─sdo2
> >                             8:226  0  14.6T  0 part
> > sdp
> >                            8:240  0  14.6T  0 disk
> > ├─sdp1
> >                             8:241  0   100M  0 part
> > /var/lib/ceph/osd/ceph-195
> > └─sdp2
> >                             8:242  0  14.6T  0 part
> > sdq
> >                           65:0    0  14.6T  0 disk
> > ├─sdq1
> >                            65:1    0   100M  0 part
> > /var/lib/ceph/osd/ceph-196
> > └─sdq2
> >                            65:2    0  14.6T  0 part
> > sdr
> >                           65:16   0  14.6T  0 disk
> > ├─sdr1
> >                            65:17   0   100M  0 part
> > /var/lib/ceph/osd/ceph-197
> > └─sdr2
> >                            65:18   0  14.6T  0 part
> > sds
> >                           65:32   0  14.6T  0 disk
> > ├─sds1
> >                            65:33   0   100M  0 part
> > /var/lib/ceph/osd/ceph-198
> > └─sds2
> >                            65:34   0  14.6T  0 part
> > sdt
> >                           65:48   0 894.3G  0 disk
> > ├─sdt1
> >                            65:49   0   100M  0 part
> > /var/lib/ceph/osd/ceph-199
> > └─sdt2
> >                            65:50   0 894.2G  0 part
> > sdu
> >                           65:64   0 894.3G  0 disk
> >
> └─ceph--e115ee52--429b--4631--92f8--a4282cb43938-osd--block--f4375a1a--6a0e--4797--a496--0b9ba0448607
> > 253:3    0 894.3G  0 lvm
> > sdv
> >                           65:80   0 894.3G  0 disk
> >
> └─ceph--bb838dc5--4eab--4461--a0cf--6439fa12706a-osd--block--f5d007a0--86b2--43ff--8750--0b24c4f823e4
> > 253:1    0 894.3G  0 lvm
> > sdw
> >                           65:96   0  14.6T  0 disk
> >
> └─ceph--d2858c65--6b29--43e8--ad4f--d92b004c2f37-osd--block--905c999e--dd4e--463c--a0e0--93a69ccc7845
> > 253:0    0  14.6T  0 lvm
> > sdx
> >                           65:112  0 223.5G  0 disk
> > ├─sdx1
> >                            65:113  0     1M  0 part
> > └─sdx2
> >                            65:114  0 223.5G  0 part
> > rbd0
> >                           252:0    0    15G  0 disk
> > rbd1
> >                           252:16   0     6T  0 disk
> > nvme0n1
> >                          259:2    0   1.5T  0 disk
> > └─md0
> >                            9:0    0   1.5T  0 raid1
> >  ├─md0p1
> >                          259:4    0    50G  0 md
> >  ├─md0p2
> >                          259:5    0    50G  0 md
> >  ├─md0p3
> >                          259:6    0    50G  0 md
> >  ├─md0p4
> >                          259:7    0    50G  0 md
> >  ├─md0p5
> >                          259:8    0    50G  0 md
> >  ├─md0p6
> >                          259:9    0    50G  0 md
> >  ├─md0p7
> >                          259:10   0    50G  0 md
> >  ├─md0p8
> >                          259:11   0    50G  0 md
> >  ├─md0p9
> >                          259:12   0    50G  0 md
> >  ├─md0p10
> >                           259:13   0    50G  0 md
> >  ├─md0p11
> >                           259:14   0    50G  0 md
> >  ├─md0p12
> >                           259:15   0    50G  0 md
> >  ├─md0p13
> >                           259:16   0    50G  0 md
> >  ├─md0p14
> >                           259:17   0    50G  0 md
> >  ├─md0p15
> >                           259:18   0    50G  0 md
> >  ├─md0p16
> >                           259:19   0    50G  0 md
> >  ├─md0p17
> >                           259:20   0    50G  0 md
> >  ├─md0p18
> >                           259:21   0    50G  0 md
> >  └─md0p19
> >                           259:22   0    50G  0 md
> > nvme1n1
> >                          259:3    0   1.5T  0 disk
> > └─md0
> >                            9:0    0   1.5T  0 raid1
> >  ├─md0p1
> >                          259:4    0    50G  0 md
> >  ├─md0p2
> >                          259:5    0    50G  0 md
> >  ├─md0p3
> >                          259:6    0    50G  0 md
> >  ├─md0p4
> >                          259:7    0    50G  0 md
> >  ├─md0p5
> >                          259:8    0    50G  0 md
> >  ├─md0p6
> >                          259:9    0    50G  0 md
> >  ├─md0p7
> >                          259:10   0    50G  0 md
> >  ├─md0p8
> >                          259:11   0    50G  0 md
> >  ├─md0p9
> >                          259:12   0    50G  0 md
> >  ├─md0p10
> >                           259:13   0    50G  0 md
> >  ├─md0p11
> >                           259:14   0    50G  0 md
> >  ├─md0p12
> >                           259:15   0    50G  0 md
> >  ├─md0p13
> >                           259:16   0    50G  0 md
> >  ├─md0p14
> >                           259:17   0    50G  0 md
> >  ├─md0p15
> >                           259:18   0    50G  0 md
> >  ├─md0p16
> >                           259:19   0    50G  0 md
> >  ├─md0p17
> >                           259:20   0    50G  0 md
> >  ├─md0p18
> >                           259:21   0    50G  0 md
> >  └─md0p19
> >                           259:22   0    50G  0 md
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@xxxxxxx
> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux