Re: Extremally need help. Openshift cluster is down :c

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On the MDS host you can see all cephadm daemon with 'cephadm ls', with 'cephadm logs --name mds.<DAEMON>' you get the logs.

Zitat von kreept.sama@xxxxxxxxx:

Sorry, dont know where to find mds logs. I just found some logs in /var/log/ceph/ceph-volume.log from mds pod and here it is(just a piece):
...
[2023-02-15 12:09:07,460][ceph_volume.main][INFO ] Running command: ceph-volume inventory --format json /dev/sda3 [2023-02-15 12:09:07,461][ceph_volume.process][INFO ] Running command: /usr/bin/lsblk -plno KNAME,NAME,TYPE [2023-02-15 12:09:07,471][ceph_volume.process][INFO ] stdout /dev/sda /dev/sda disk [2023-02-15 12:09:07,472][ceph_volume.process][INFO ] stdout /dev/sda1 /dev/sda1 part [2023-02-15 12:09:07,472][ceph_volume.process][INFO ] stdout /dev/sda2 /dev/sda2 part [2023-02-15 12:09:07,472][ceph_volume.process][INFO ] stdout /dev/sda3 /dev/sda3 part [2023-02-15 12:09:07,472][ceph_volume.process][INFO ] stdout /dev/sda4 /dev/sda4 part [2023-02-15 12:09:07,472][ceph_volume.process][INFO ] stdout /dev/sdb /dev/sdb disk [2023-02-15 12:09:07,472][ceph_volume.process][INFO ] stdout /dev/sr0 /dev/sr0 rom [2023-02-15 12:09:07,472][ceph_volume.process][INFO ] stdout /dev/dm-0 /dev/mapper/ceph--893e62b0--dc66--4db6--9487--404cad5907d6-osd--block--19ef4373--d828--4259--9159--86f2e89f4397 lvm [2023-02-15 12:09:07,479][ceph_volume.process][INFO ] Running command: /usr/sbin/lvs --noheadings --readonly --separator=";" -a --units=b --nosuffix -S lv_path=/dev/sda3 -o lv_tags,lv_path,lv_name,vg_name,lv_uuid,lv_size [2023-02-15 12:09:07,525][ceph_volume.process][INFO ] stderr Udev is running and DM_DISABLE_UDEV environment variable is set. Bypassing udev, LVM will manage logical volume symlinks in device directory. [2023-02-15 12:09:07,526][ceph_volume.process][INFO ] stderr Udev is running and DM_DISABLE_UDEV environment variable is set. Bypassing udev, LVM will obtain device list by scanning device directory. [2023-02-15 12:09:07,526][ceph_volume.process][INFO ] Running command: /usr/bin/lsblk --nodeps -P -o NAME,KNAME,MAJ:MIN,FSTYPE,MOUNTPOINT,LABEL,UUID,RO,RM,MODEL,SIZE,STATE,OWNER,GROUP,MODE,ALIGNMENT,PHY-SEC,LOG-SEC,ROTA,SCHED,TYPE,DISC-ALN,DISC-GRAN,DISC-MAX,DISC-ZERO,PKNAME,PARTLABEL /dev/sda3 [2023-02-15 12:09:07,544][ceph_volume.process][INFO ] stdout NAME="sda3" KNAME="sda3" MAJ:MIN="8:3" FSTYPE="" MOUNTPOINT="" LABEL="" UUID="" RO="0" RM="0" MODEL="" SIZE="1M" STATE="" OWNER="root" GROUP="disk" MODE="brw-rw----" ALIGNMENT="0" PHY-SEC="512" LOG-SEC="512" ROTA="1" SCHED="bfq" TYPE="part" DISC-ALN="0" DISC-GRAN="4K" DISC-MAX="1G" DISC-ZERO="0" PKNAME="sda" PARTLABEL="BIOS-BOOT" [2023-02-15 12:09:07,544][ceph_volume.process][INFO ] Running command: /usr/sbin/blkid -c /dev/null -p /dev/sda3 [2023-02-15 12:09:07,553][ceph_volume.process][INFO ] stdout /dev/sda3: PART_ENTRY_SCHEME="gpt" PART_ENTRY_NAME="BIOS-BOOT" PART_ENTRY_UUID="860433f4-8158-432c-88aa-e42901086299" PART_ENTRY_TYPE="21686148-6449-6e6f-744e-656564454649" PART_ENTRY_NUMBER="3" PART_ENTRY_OFFSET="1048576" PART_ENTRY_SIZE="2048" PART_ENTRY_DISK="8:0" [2023-02-15 12:09:07,554][ceph_volume.process][INFO ] Running command: /usr/sbin/pvs --noheadings --readonly --units=b --nosuffix --separator=";" -o vg_name,pv_count,lv_count,vg_attr,vg_extent_count,vg_free_count,vg_extent_size /dev/sda3 [2023-02-15 12:09:07,595][ceph_volume.process][INFO ] stderr Udev is running and DM_DISABLE_UDEV environment variable is set. Bypassing udev, LVM will manage logical volume symlinks in device directory. [2023-02-15 12:09:07,595][ceph_volume.process][INFO ] stderr Udev is running and DM_DISABLE_UDEV environment variable is set. Bypassing udev, LVM will obtain device list by scanning device directory. [2023-02-15 12:09:07,596][ceph_volume.process][INFO ] stderr Failed to find device for physical volume "/dev/sda3". [2023-02-15 12:09:07,596][ceph_volume.process][INFO ] Running command: /usr/bin/ceph-bluestore-tool show-label --dev /dev/sda3 [2023-02-15 12:09:07,630][ceph_volume.process][INFO ] stderr unable to read label for /dev/sda3: (2) No such file or directory [2023-02-15 12:09:07,630][ceph_volume.process][INFO ] Running command: /usr/bin/ceph-bluestore-tool show-label --dev /dev/sda3 [2023-02-15 12:09:07,672][ceph_volume.process][INFO ] stderr unable to read label for /dev/sda3: (2) No such file or directory [2023-02-15 12:09:07,673][ceph_volume.process][INFO ] Running command: /usr/sbin/udevadm info --query=property /dev/sda3 [2023-02-15 12:09:07,687][ceph_volume.process][INFO ] stdout DEVLINKS=/dev/disk/by-partuuid/860433f4-8158-432c-88aa-e42901086299 /dev/disk/by-id/scsi-SQEMU_QEMU_HARDDISK_6ef7d571-c0dc-4394-bcc0-4625a857778b-part3 /dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_6ef7d571-c0dc-4394-bcc0-4625a857778b-part3 /dev/disk/by-path/pci-0000:00:06.0-scsi-0:0:0:0-part3 /dev/disk/by-partlabel/BIOS-BOOT [2023-02-15 12:09:07,688][ceph_volume.process][INFO ] stdout DEVNAME=/dev/sda3 [2023-02-15 12:09:07,688][ceph_volume.process][INFO ] stdout DEVPATH=/devices/pci0000:00/0000:00:06.0/virtio2/host2/target2:0:0/2:0:0:0/block/sda/sda3 [2023-02-15 12:09:07,688][ceph_volume.process][INFO ] stdout DEVTYPE=partition
[2023-02-15 12:09:07,688][ceph_volume.process][INFO  ] stdout ID_BUS=scsi
[2023-02-15 12:09:07,688][ceph_volume.process][INFO ] stdout ID_MODEL=QEMU_HARDDISK [2023-02-15 12:09:07,688][ceph_volume.process][INFO ] stdout ID_MODEL_ENC=QEMU\x20HARDDISK\x20\x20\x20 [2023-02-15 12:09:07,688][ceph_volume.process][INFO ] stdout ID_PART_ENTRY_DISK=8:0 [2023-02-15 12:09:07,688][ceph_volume.process][INFO ] stdout ID_PART_ENTRY_NAME=BIOS-BOOT
....

Also after some manipulation our filesystem now fine(few steps later we create new metadata pool and restore current from new). But mds still read only. Our ceph shows only this issue :
  cluster:
    id:     9213604e-b0b6-49d5-bcb3-f55ab3d79119
    health: HEALTH_WARN
            1 MDSs are read only
            21 daemons have recently crashed

  services:
    mon: 5 daemons, quorum bd,bj,bm,bn,bo (age 8h)
    mgr: a(active, since 102m)
    mds: 1/1 daemons up, 1 hot standby
    osd: 9 osds: 9 up (since 8h), 9 in (since 7d)
    rgw: 1 daemon active (1 hosts, 1 zones)

Thank you
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx


_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux