Unfortunately nothing.... Is there any way to make it more verbose? On 14.02.22, 11:48, "Eugen Block" <eblock@xxxxxx> wrote: What does 'dmesg' reveal? Zitat von Lo Re Giuseppe <giuseppe.lore@xxxxxxx>: > root@fulen-w006:~# ll client.fulen.keyring > -rw-r--r-- 1 root root 69 Feb 11 15:30 client.fulen.keyring > root@fulen-w006:~# ll ceph.conf > -rw-r--r-- 1 root root 118 Feb 11 19:15 ceph.conf > root@fulen-w006:~# rbd -c ceph.conf --id fulen --keyring > client.fulen.keyring map fulen-nvme-meta/test-loreg-3 > rbd: sysfs write failed > In some cases useful info is found in syslog - try "dmesg | tail". > rbd: map failed: (1) Operation not permitted > > where test-loreg-3 is an image in a EC pool. > > root@fulen-w006:~# rbd -c ceph.conf --id fulen --keyring > client.fulen.keyring info fulen-nvme-meta/test-loreg-3 > rbd image 'test-loreg-3': > size 1 GiB in 256 objects > order 22 (4 MiB objects) > snapshot_count: 0 > id: e2375bbddf414a > data_pool: fulen-hdd-data > block_name_prefix: rbd_data.36.e2375bbddf414a > format: 2 > features: layering, exclusive-lock, data-pool > op_features: > flags: > create_timestamp: Thu Feb 10 18:17:42 2022 > access_timestamp: Thu Feb 10 18:17:42 2022 > modify_timestamp: Thu Feb 10 18:17:42 2022 > > Giuseppe > > On 11.02.22, 14:52, "Eugen Block" <eblock@xxxxxx> wrote: > > How are the permissions of the client keyring on both systems? > > Zitat von Lo Re Giuseppe <giuseppe.lore@xxxxxxx>: > > > Hi, > > > > It's a single ceph cluster, I'm testing from 2 different client nodes. > > The caps are below. > > I think is unlikely that caps are the cause as they work from one > > client node, same ceph user, and not from the other one... > > > > Cheers, > > > > Giuseppe > > > > > > [root@naret-monitor01 ~]# ceph auth get client.fulen > > exported keyring for client.fulen > > [client.fulen] > > key = xxxxxxxx > > caps mgr = "profile rbd pool=fulen-hdd, profile rbd > > pool=fulen-nvme, profile rbd pool=fulen-dcache, profile rbd > > pool=fulen-dcache-data, profile rbd pool=fulen-dcache-meta, profile > > rbd pool=fulen-hdd-data, profile rbd pool=fulen-nvme-meta" > > caps mon = "profile rbd" > > caps osd = "profile rbd pool=fulen-hdd, profile rbd > > pool=fulen-nvme, profile rbd pool=fulen-dcache, profile rbd > > pool=fulen-dcache-data, profile rbd pool=fulen-dcache-meta, profile > > rbd pool=fulen-hdd-data, profile rbd pool=fulen-nvme-meta" > > > > > > > > On 11.02.22, 13:22, "Eugen Block" <eblock@xxxxxx> wrote: > > > > Hi, > > > > the first thing coming to mind are the user's caps. Which > permissions > > do they have? Have you compared 'ceph auth get > client.fulen' on both > > clusters? Please paste the output from both clusters and redact > > sensitive information. > > > > > > Zitat von Lo Re Giuseppe <giuseppe.lore@xxxxxxx>: > > > > > Hi all, > > > > > > This is my first post to this user group, I’m not a ceph expert, > > > sorry if I say/ask anything trivial. > > > > > > On a Kubernetes cluster I have an issue in creating > volumes from a > > > (csi) ceph EC pool. > > > > > > I can reproduce the problem from rbd cli like this from > one of the > > > k8s worker nodes: > > > > > > “”” > > > root@fulen-w006:~# ceph -v > > > ceph version 15.2.14 (cd3bb7e87a2f62c1b862ff3fd8b1eec13391a5be) > > > octopus (stable) > > > > > > root@fulen-w006:~# rbd -m 148.187.20.141:6789 --id fulen > --keyfile > > > key create test-loreg-3 --data-pool fulen-hdd-data --pool > > > fulen-nvme-meta --size 1G > > > > > > root@fulen-w006:~# rbd -m 148.187.20.141:6789 --id fulen > --keyfile > > > key map fulen-nvme-meta/test-loreg-3 > > > ... > > > rbd: sysfs write failed > > > ... > > > In some cases useful info is found in syslog - try > "dmesg | tail". > > > rbd: map failed: (1) Operation not permitted > > > “”” > > > > > > The same sequence of operations works on a different > node (not part > > > of the k8s cluster, completely different setup): > > > > > > “”” > > > root@storagesmw: # ceph -v > > > ceph version 15.2.13 (c44bc49e7a57a87d84dfff2a077a2058aa2172e2) > > > octopus (stable) > > > > > > root@storagesmw: # rbd -m 148.187.20.141:6789 --id fulen > --keyfile > > > client.fulen.key create test-loreg-4 --data-pool fulen-hdd-data > > > --pool fulen-nvme-meta --size 1G > > > > > > root@storagesmw: # rbd -m 148.187.20.141:6789 --id fulen > --keyfile > > > client.fulen.key info fulen-nvme-meta/test-loreg-4 > > > rbd image 'test-loreg-4': > > > size 1 GiB in 256 objects > > > order 22 (4 MiB objects) > > > snapshot_count: 0 > > > id: cafc436ff3573 > > > data_pool: fulen-hdd-data > > > block_name_prefix: rbd_data.36.cafc436ff3573 > > > format: 2 > > > features: layering, exclusive-lock, object-map, > fast-diff, > > > deep-flatten, data-pool > > > op_features: > > > flags: > > > create_timestamp: Thu Feb 10 18:23:26 2022 > > > access_timestamp: Thu Feb 10 18:23:26 2022 > > > modify_timestamp: Thu Feb 10 18:23:26 2022 > > > > > > root@storagesmw: # rbd -m 148.187.20.141:6789 --id fulen > --keyfile > > > client.fulen.key map fulen-nvme-meta/test-loreg-4 > > > RBD image feature set mismatch. You can disable features > unsupported > > > by the kernel with "rbd feature disable > fulen-nvme-meta/test-loreg-4 > > > object-map fast-diff deep-flatten". > > > In some cases useful info is found in syslog - try > "dmesg | tail". > > > rbd: map failed: (6) No such device or address > > > > > > root@storagesmw: # rbd -m 148.187.20.141:6789 --id fulen > --keyfile > > > client.fulen.key feature disable fulen-nvme-meta/test-loreg-4 > > > object-map fast-diff deep-flatten > > > > > > root@storagesmw: # rbd -m 148.187.20.141:6789 --id fulen > --keyfile > > > client.fulen.key map fulen-nvme-meta/test-loreg-4 > > > /dev/rbd0 > > > “”” > > > > > > The 2 nodes OS release and kernel are below. > > > > > > Does anyone have any advice on how to debug this? > > > > > > Thanks in advance, > > > > > > Giuseppe > > > > > > ============================================== Fulen-w006: > > > > > > root@fulen-w006:~# cat /etc/os-release > > > NAME="Ubuntu" > > > VERSION="20.04.3 LTS (Focal Fossa)" > > > ID=ubuntu > > > ID_LIKE=debian > > > PRETTY_NAME="Ubuntu 20.04.3 LTS" > > > VERSION_ID="20.04" > > > HOME_URL="https://www.ubuntu.com/" > > > SUPPORT_URL="https://help.ubuntu.com/" > > > BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" > > > > > > PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" > > > VERSION_CODENAME=focal > > > UBUNTU_CODENAME=focal > > > root@fulen-w006:~# uname -a > > > Linux fulen-w006.cscs.ch 5.4.0-96-generic #109-Ubuntu > SMP Wed Jan 12 > > > 16:49:16 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux > > > > > > Storagesmw: > > > > > > root@storagesmw:~/loreg/ceph_conf # cat /etc/os-release > > > NAME="Red Hat Enterprise Linux Server" > > > VERSION="7.8 (Maipo)" > > > ID="rhel" > > > ID_LIKE="fedora" > > > VARIANT="Server" > > > VARIANT_ID="server" > > > VERSION_ID="7.8" > > > PRETTY_NAME="Red Hat Enterprise Linux" > > > ANSI_COLOR="0;31" > > > CPE_NAME="cpe:/o:redhat:enterprise_linux:7.8:GA:server" > > > HOME_URL="https://www.redhat.com/" > > > BUG_REPORT_URL="https://bugzilla.redhat.com/" > > > > > > REDHAT_BUGZILLA_PRODUCT="Red Hat Enterprise Linux 7" > > > REDHAT_BUGZILLA_PRODUCT_VERSION=7.8 > > > REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux" > > > REDHAT_SUPPORT_PRODUCT_VERSION="7.8" > > > root@storagesmw:~/loreg/ceph_conf # uname -a > > > Linux storagesmw.cscs.ch 3.10.0-1127.19.1.el7.x86_64 #1 > SMP Tue Aug > > > 11 19:12:04 EDT 2020 x86_64 x86_64 x86_64 GNU/Linux > > > > > > > > > > > > > > > > > > _______________________________________________ > > > ceph-users mailing list -- ceph-users@xxxxxxx > > > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > > > > > > > _______________________________________________ > > ceph-users mailing list -- ceph-users@xxxxxxx > > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx