Re: RBD map issue

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Unfortunately nothing....
Is there any way to make it more verbose?

On 14.02.22, 11:48, "Eugen Block" <eblock@xxxxxx> wrote:

    What does 'dmesg' reveal?


    Zitat von Lo Re  Giuseppe <giuseppe.lore@xxxxxxx>:

    > root@fulen-w006:~# ll client.fulen.keyring
    > -rw-r--r-- 1 root root 69 Feb 11 15:30 client.fulen.keyring
    > root@fulen-w006:~# ll ceph.conf
    > -rw-r--r-- 1 root root 118 Feb 11 19:15 ceph.conf
    > root@fulen-w006:~# rbd -c ceph.conf --id fulen --keyring  
    > client.fulen.keyring map fulen-nvme-meta/test-loreg-3
    > rbd: sysfs write failed
    > In some cases useful info is found in syslog - try "dmesg | tail".
    > rbd: map failed: (1) Operation not permitted
    >
    > where test-loreg-3 is an image in a EC pool.
    >
    > root@fulen-w006:~# rbd -c ceph.conf --id fulen --keyring  
    > client.fulen.keyring info fulen-nvme-meta/test-loreg-3
    > rbd image 'test-loreg-3':
    > 	size 1 GiB in 256 objects
    > 	order 22 (4 MiB objects)
    > 	snapshot_count: 0
    > 	id: e2375bbddf414a
    > 	data_pool: fulen-hdd-data
    > 	block_name_prefix: rbd_data.36.e2375bbddf414a
    > 	format: 2
    > 	features: layering, exclusive-lock, data-pool
    > 	op_features:
    > 	flags:
    > 	create_timestamp: Thu Feb 10 18:17:42 2022
    > 	access_timestamp: Thu Feb 10 18:17:42 2022
    > 	modify_timestamp: Thu Feb 10 18:17:42 2022
    >
    > Giuseppe
    >
    > On 11.02.22, 14:52, "Eugen Block" <eblock@xxxxxx> wrote:
    >
    >     How are the permissions of the client keyring on both systems?
    >
    >     Zitat von Lo Re  Giuseppe <giuseppe.lore@xxxxxxx>:
    >
    >     > Hi,
    >     >
    >     > It's a single ceph cluster, I'm testing from 2 different client nodes.
    >     > The caps are below.
    >     > I think is unlikely that caps are the cause as they work from one
    >     > client node, same ceph user, and not from the other one...
    >     >
    >     > Cheers,
    >     >
    >     > Giuseppe
    >     >
    >     >
    >     > [root@naret-monitor01 ~]# ceph auth get client.fulen
    >     > exported keyring for client.fulen
    >     > [client.fulen]
    >     > 	key = xxxxxxxx
    >     > 	caps mgr = "profile rbd pool=fulen-hdd, profile rbd
    >     > pool=fulen-nvme, profile rbd pool=fulen-dcache, profile rbd
    >     > pool=fulen-dcache-data, profile rbd pool=fulen-dcache-meta, profile
    >     > rbd pool=fulen-hdd-data, profile rbd pool=fulen-nvme-meta"
    >     > 	caps mon = "profile rbd"
    >     > 	caps osd = "profile rbd pool=fulen-hdd, profile rbd
    >     > pool=fulen-nvme, profile rbd pool=fulen-dcache, profile rbd
    >     > pool=fulen-dcache-data, profile rbd pool=fulen-dcache-meta, profile
    >     > rbd pool=fulen-hdd-data, profile rbd pool=fulen-nvme-meta"
    >     >
    >     >
    >     >
    >     > On 11.02.22, 13:22, "Eugen Block" <eblock@xxxxxx> wrote:
    >     >
    >     >     Hi,
    >     >
    >     >     the first thing coming to mind are the user's caps. Which  
    > permissions
    >     >     do they have? Have you compared 'ceph auth get  
    > client.fulen' on both
    >     >     clusters? Please paste the output from both clusters and redact
    >     >     sensitive information.
    >     >
    >     >
    >     >     Zitat von Lo Re  Giuseppe <giuseppe.lore@xxxxxxx>:
    >     >
    >     >     > Hi all,
    >     >     >
    >     >     > This is my first post to this user group, I’m not a ceph expert,
    >     >     > sorry if I say/ask anything trivial.
    >     >     >
    >     >     > On a Kubernetes cluster I have an issue in creating  
    > volumes from a
    >     >     > (csi) ceph EC pool.
    >     >     >
    >     >     > I can reproduce the problem from rbd cli like this from  
    > one of the
    >     >     > k8s worker nodes:
    >     >     >
    >     >     > “””
    >     >     > root@fulen-w006:~# ceph -v
    >     >     > ceph version 15.2.14 (cd3bb7e87a2f62c1b862ff3fd8b1eec13391a5be)
    >     >     > octopus (stable)
    >     >     >
    >     >     > root@fulen-w006:~# rbd -m 148.187.20.141:6789 --id fulen  
    > --keyfile
    >     >     > key create test-loreg-3 --data-pool fulen-hdd-data --pool
    >     >     > fulen-nvme-meta --size 1G
    >     >     >
    >     >     > root@fulen-w006:~# rbd -m 148.187.20.141:6789 --id fulen  
    > --keyfile
    >     >     > key map fulen-nvme-meta/test-loreg-3
    >     >     > ...
    >     >     > rbd: sysfs write failed
    >     >     > ...
    >     >     > In some cases useful info is found in syslog - try  
    > "dmesg | tail".
    >     >     > rbd: map failed: (1) Operation not permitted
    >     >     > “””
    >     >     >
    >     >     > The same sequence of operations works on a different  
    > node (not part
    >     >     > of the k8s cluster, completely different setup):
    >     >     >
    >     >     > “””
    >     >     > root@storagesmw: # ceph -v
    >     >     > ceph version 15.2.13 (c44bc49e7a57a87d84dfff2a077a2058aa2172e2)
    >     >     > octopus (stable)
    >     >     >
    >     >     > root@storagesmw: # rbd -m 148.187.20.141:6789 --id fulen  
    > --keyfile
    >     >     > client.fulen.key create test-loreg-4 --data-pool fulen-hdd-data
    >     >     > --pool fulen-nvme-meta --size 1G
    >     >     >
    >     >     > root@storagesmw: # rbd -m 148.187.20.141:6789 --id fulen  
    > --keyfile
    >     >     > client.fulen.key info fulen-nvme-meta/test-loreg-4
    >     >     > rbd image 'test-loreg-4':
    >     >     >         size 1 GiB in 256 objects
    >     >     >         order 22 (4 MiB objects)
    >     >     >         snapshot_count: 0
    >     >     >         id: cafc436ff3573
    >     >     >         data_pool: fulen-hdd-data
    >     >     >         block_name_prefix: rbd_data.36.cafc436ff3573
    >     >     >         format: 2
    >     >     >         features: layering, exclusive-lock, object-map,  
    > fast-diff,
    >     >     > deep-flatten, data-pool
    >     >     >         op_features:
    >     >     >         flags:
    >     >     >         create_timestamp: Thu Feb 10 18:23:26 2022
    >     >     >         access_timestamp: Thu Feb 10 18:23:26 2022
    >     >     >         modify_timestamp: Thu Feb 10 18:23:26 2022
    >     >     >
    >     >     > root@storagesmw: # rbd -m 148.187.20.141:6789 --id fulen  
    > --keyfile
    >     >     > client.fulen.key map fulen-nvme-meta/test-loreg-4
    >     >     > RBD image feature set mismatch. You can disable features  
    > unsupported
    >     >     > by the kernel with "rbd feature disable  
    > fulen-nvme-meta/test-loreg-4
    >     >     > object-map fast-diff deep-flatten".
    >     >     > In some cases useful info is found in syslog - try  
    > "dmesg | tail".
    >     >     > rbd: map failed: (6) No such device or address
    >     >     >
    >     >     > root@storagesmw: # rbd -m 148.187.20.141:6789 --id fulen  
    > --keyfile
    >     >     > client.fulen.key feature disable fulen-nvme-meta/test-loreg-4
    >     >     > object-map fast-diff deep-flatten
    >     >     >
    >     >     > root@storagesmw: # rbd -m 148.187.20.141:6789 --id fulen  
    > --keyfile
    >     >     > client.fulen.key map fulen-nvme-meta/test-loreg-4
    >     >     > /dev/rbd0
    >     >     > “””
    >     >     >
    >     >     > The 2 nodes OS release and kernel are below.
    >     >     >
    >     >     > Does anyone have any advice on how to debug this?
    >     >     >
    >     >     > Thanks in advance,
    >     >     >
    >     >     > Giuseppe
    >     >     >
    >     >     > ============================================== Fulen-w006:
    >     >     >
    >     >     > root@fulen-w006:~# cat /etc/os-release
    >     >     > NAME="Ubuntu"
    >     >     > VERSION="20.04.3 LTS (Focal Fossa)"
    >     >     > ID=ubuntu
    >     >     > ID_LIKE=debian
    >     >     > PRETTY_NAME="Ubuntu 20.04.3 LTS"
    >     >     > VERSION_ID="20.04"
    >     >     > HOME_URL="https://www.ubuntu.com/";
    >     >     > SUPPORT_URL="https://help.ubuntu.com/";
    >     >     > BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/";
    >     >     >
    >     >  
    > PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy";
    >     >     > VERSION_CODENAME=focal
    >     >     > UBUNTU_CODENAME=focal
    >     >     > root@fulen-w006:~# uname -a
    >     >     > Linux fulen-w006.cscs.ch 5.4.0-96-generic #109-Ubuntu  
    > SMP Wed Jan 12
    >     >     > 16:49:16 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux
    >     >     >
    >     >     > Storagesmw:
    >     >     >
    >     >     > root@storagesmw:~/loreg/ceph_conf # cat /etc/os-release
    >     >     > NAME="Red Hat Enterprise Linux Server"
    >     >     > VERSION="7.8 (Maipo)"
    >     >     > ID="rhel"
    >     >     > ID_LIKE="fedora"
    >     >     > VARIANT="Server"
    >     >     > VARIANT_ID="server"
    >     >     > VERSION_ID="7.8"
    >     >     > PRETTY_NAME="Red Hat Enterprise Linux"
    >     >     > ANSI_COLOR="0;31"
    >     >     > CPE_NAME="cpe:/o:redhat:enterprise_linux:7.8:GA:server"
    >     >     > HOME_URL="https://www.redhat.com/";
    >     >     > BUG_REPORT_URL="https://bugzilla.redhat.com/";
    >     >     >
    >     >     > REDHAT_BUGZILLA_PRODUCT="Red Hat Enterprise Linux 7"
    >     >     > REDHAT_BUGZILLA_PRODUCT_VERSION=7.8
    >     >     > REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux"
    >     >     > REDHAT_SUPPORT_PRODUCT_VERSION="7.8"
    >     >     > root@storagesmw:~/loreg/ceph_conf # uname -a
    >     >     > Linux storagesmw.cscs.ch 3.10.0-1127.19.1.el7.x86_64 #1  
    > SMP Tue Aug
    >     >     > 11 19:12:04 EDT 2020 x86_64 x86_64 x86_64 GNU/Linux
    >     >     >
    >     >     >
    >     >     >
    >     >     >
    >     >     >
    >     >     > _______________________________________________
    >     >     > ceph-users mailing list -- ceph-users@xxxxxxx
    >     >     > To unsubscribe send an email to ceph-users-leave@xxxxxxx
    >     >
    >     >
    >     >
    >     >     _______________________________________________
    >     >     ceph-users mailing list -- ceph-users@xxxxxxx
    >     >     To unsubscribe send an email to ceph-users-leave@xxxxxxx




_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux