Sorry, dont know where to find mds logs. I just found some logs in /var/log/ceph/ceph-volume.log from mds pod and here it is(just a piece): ... [2023-02-15 12:09:07,460][ceph_volume.main][INFO ] Running command: ceph-volume inventory --format json /dev/sda3 [2023-02-15 12:09:07,461][ceph_volume.process][INFO ] Running command: /usr/bin/lsblk -plno KNAME,NAME,TYPE [2023-02-15 12:09:07,471][ceph_volume.process][INFO ] stdout /dev/sda /dev/sda disk [2023-02-15 12:09:07,472][ceph_volume.process][INFO ] stdout /dev/sda1 /dev/sda1 part [2023-02-15 12:09:07,472][ceph_volume.process][INFO ] stdout /dev/sda2 /dev/sda2 part [2023-02-15 12:09:07,472][ceph_volume.process][INFO ] stdout /dev/sda3 /dev/sda3 part [2023-02-15 12:09:07,472][ceph_volume.process][INFO ] stdout /dev/sda4 /dev/sda4 part [2023-02-15 12:09:07,472][ceph_volume.process][INFO ] stdout /dev/sdb /dev/sdb disk [2023-02-15 12:09:07,472][ceph_volume.process][INFO ] stdout /dev/sr0 /dev/sr0 rom [2023-02-15 12:09:07,472][ceph_volume.process][INFO ] stdout /dev/dm-0 /dev/mapper/ceph--893e62b0--dc66--4db6--9487--404cad5907d6-osd--block--19ef4373--d828--4259--9159--86f2e89f4397 lvm [2023-02-15 12:09:07,479][ceph_volume.process][INFO ] Running command: /usr/sbin/lvs --noheadings --readonly --separator=";" -a --units=b --nosuffix -S lv_path=/dev/sda3 -o lv_tags,lv_path,lv_name,vg_name,lv_uuid,lv_size [2023-02-15 12:09:07,525][ceph_volume.process][INFO ] stderr Udev is running and DM_DISABLE_UDEV environment variable is set. Bypassing udev, LVM will manage logical volume symlinks in device directory. [2023-02-15 12:09:07,526][ceph_volume.process][INFO ] stderr Udev is running and DM_DISABLE_UDEV environment variable is set. Bypassing udev, LVM will obtain device list by scanning device directory. [2023-02-15 12:09:07,526][ceph_volume.process][INFO ] Running command: /usr/bin/lsblk --nodeps -P -o NAME,KNAME,MAJ:MIN,FSTYPE,MOUNTPOINT,LABEL,UUID,RO,RM,MODEL,SIZE,STATE,OWNER,GROUP,MODE,ALIGNMENT,PHY-SEC,LOG-SEC,ROTA,SCHED,TYPE,DISC-ALN,DISC-GRAN,DISC-MAX,DISC-ZERO,PKNAME,PARTLABEL /dev/sda3 [2023-02-15 12:09:07,544][ceph_volume.process][INFO ] stdout NAME="sda3" KNAME="sda3" MAJ:MIN="8:3" FSTYPE="" MOUNTPOINT="" LABEL="" UUID="" RO="0" RM="0" MODEL="" SIZE="1M" STATE="" OWNER="root" GROUP="disk" MODE="brw-rw----" ALIGNMENT="0" PHY-SEC="512" LOG-SEC="512" ROTA="1" SCHED="bfq" TYPE="part" DISC-ALN="0" DISC-GRAN="4K" DISC-MAX="1G" DISC-ZERO="0" PKNAME="sda" PARTLABEL="BIOS-BOOT" [2023-02-15 12:09:07,544][ceph_volume.process][INFO ] Running command: /usr/sbin/blkid -c /dev/null -p /dev/sda3 [2023-02-15 12:09:07,553][ceph_volume.process][INFO ] stdout /dev/sda3: PART_ENTRY_SCHEME="gpt" PART_ENTRY_NAME="BIOS-BOOT" PART_ENTRY_UUID="860433f4-8158-432c-88aa-e42901086299" PART_ENTRY_TYPE="21686148-6449-6e6f-744e-656564454649" PART_ENTRY_NUMBER="3" PART_ENTRY_OFFSET="1048576" PART_ENTRY_SIZE="2048" PART_ENTRY_DISK="8:0" [2023-02-15 12:09:07,554][ceph_volume.process][INFO ] Running command: /usr/sbin/pvs --noheadings --readonly --units=b --nosuffix --separator=";" -o vg_name,pv_count,lv_count,vg_attr,vg_extent_count,vg_free_count,vg_extent_size /dev/sda3 [2023-02-15 12:09:07,595][ceph_volume.process][INFO ] stderr Udev is running and DM_DISABLE_UDEV environment variable is set. Bypassing udev, LVM will manage logical volume symlinks in device directory. [2023-02-15 12:09:07,595][ceph_volume.process][INFO ] stderr Udev is running and DM_DISABLE_UDEV environment variable is set. Bypassing udev, LVM will obtain device list by scanning device directory. [2023-02-15 12:09:07,596][ceph_volume.process][INFO ] stderr Failed to find device for physical volume "/dev/sda3". [2023-02-15 12:09:07,596][ceph_volume.process][INFO ] Running command: /usr/bin/ceph-bluestore-tool show-label --dev /dev/sda3 [2023-02-15 12:09:07,630][ceph_volume.process][INFO ] stderr unable to read label for /dev/sda3: (2) No such file or directory [2023-02-15 12:09:07,630][ceph_volume.process][INFO ] Running command: /usr/bin/ceph-bluestore-tool show-label --dev /dev/sda3 [2023-02-15 12:09:07,672][ceph_volume.process][INFO ] stderr unable to read label for /dev/sda3: (2) No such file or directory [2023-02-15 12:09:07,673][ceph_volume.process][INFO ] Running command: /usr/sbin/udevadm info --query=property /dev/sda3 [2023-02-15 12:09:07,687][ceph_volume.process][INFO ] stdout DEVLINKS=/dev/disk/by-partuuid/860433f4-8158-432c-88aa-e42901086299 /dev/disk/by-id/scsi-SQEMU_QEMU_HARDDISK_6ef7d571-c0dc-4394-bcc0-4625a857778b-part3 /dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_6ef7d571-c0dc-4394-bcc0-4625a857778b-part3 /dev/disk/by-path/pci-0000:00:06.0-scsi-0:0:0:0-part3 /dev/disk/by-partlabel/BIOS-BOOT [2023-02-15 12:09:07,688][ceph_volume.process][INFO ] stdout DEVNAME=/dev/sda3 [2023-02-15 12:09:07,688][ceph_volume.process][INFO ] stdout DEVPATH=/devices/pci0000:00/0000:00:06.0/virtio2/host2/target2:0:0/2:0:0:0/block/sda/sda3 [2023-02-15 12:09:07,688][ceph_volume.process][INFO ] stdout DEVTYPE=partition [2023-02-15 12:09:07,688][ceph_volume.process][INFO ] stdout ID_BUS=scsi [2023-02-15 12:09:07,688][ceph_volume.process][INFO ] stdout ID_MODEL=QEMU_HARDDISK [2023-02-15 12:09:07,688][ceph_volume.process][INFO ] stdout ID_MODEL_ENC=QEMU\x20HARDDISK\x20\x20\x20 [2023-02-15 12:09:07,688][ceph_volume.process][INFO ] stdout ID_PART_ENTRY_DISK=8:0 [2023-02-15 12:09:07,688][ceph_volume.process][INFO ] stdout ID_PART_ENTRY_NAME=BIOS-BOOT .... Also after some manipulation our filesystem now fine(few steps later we create new metadata pool and restore current from new). But mds still read only. Our ceph shows only this issue : cluster: id: 9213604e-b0b6-49d5-bcb3-f55ab3d79119 health: HEALTH_WARN 1 MDSs are read only 21 daemons have recently crashed services: mon: 5 daemons, quorum bd,bj,bm,bn,bo (age 8h) mgr: a(active, since 102m) mds: 1/1 daemons up, 1 hot standby osd: 9 osds: 9 up (since 8h), 9 in (since 7d) rgw: 1 daemon active (1 hosts, 1 zones) Thank you _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx