Hi . Any input will be of great help. Thanks once again. Lokendra On Fri, 14 Apr, 2023, 3:47 pm Lokendra Rathour, <lokendrarathour@xxxxxxxxx> wrote: > Hi Team, > their is one additional observation. > Mount as the client is working fine from one of the Ceph nodes. > Command *: sudo mount -t ceph :/ /mnt/imgs -o > name=foo,secret=AQABDzRkTaJCEhAAC7rC6E68ofwdfULnx6qX/VDA== * > > *we are not passing the Monitor address, instead, DNS SRV is configured as > per:* > https://docs.ceph.com/en/quincy/rados/configuration/mon-lookup-dns/ > > mount works fine in this case. > > #### > > But if we try to mount from the other Location i.e from another > VM/client(non-Ceph Node) > we are getting the error : > mount -t ceph :/ /mnt/imgs -o > name=foo,secret=AQABDzRkTaJCEhAAC7rC6E68ofwULnx6qX/VDA== -v > *mount: /mnt/image: mount point does not exist.* > > the document says that if we do not pass the monitor address, it tries > discovering the monitor address from DNS Servers, but in actual it is not > happening. > > > > On Tue, Apr 11, 2023 at 6:48 PM Lokendra Rathour < > lokendrarathour@xxxxxxxxx> wrote: > >> Ceph version Quincy. >> >> But now I am able to resolve the issue. >> >> During mount i will not pass any monitor details, it will be >> auto-discovered via SRV. >> >> On Tue, Apr 11, 2023 at 6:09 PM Eugen Block <eblock@xxxxxx> wrote: >> >>> What ceph version is this? Could it be this bug [1]? Although the >>> error message is different, not sure if it could be the same issue, >>> and I don't have anything to test ipv6 with. >>> >>> [1] https://tracker.ceph.com/issues/47300 >>> >>> Zitat von Lokendra Rathour <lokendrarathour@xxxxxxxxx>: >>> >>> > Hi All, >>> > Requesting any inputs around the issue raised. >>> > >>> > Best Regards, >>> > Lokendra >>> > >>> > On Tue, 24 Jan, 2023, 7:32 pm Lokendra Rathour, < >>> lokendrarathour@xxxxxxxxx> >>> > wrote: >>> > >>> >> Hi Team, >>> >> >>> >> >>> >> >>> >> We have a ceph cluster with 3 storage nodes: >>> >> >>> >> 1. storagenode1 - abcd:abcd:abcd::21 >>> >> >>> >> 2. storagenode2 - abcd:abcd:abcd::22 >>> >> >>> >> 3. storagenode3 - abcd:abcd:abcd::23 >>> >> >>> >> >>> >> >>> >> The requirement is to mount ceph using the domain name of MON node: >>> >> >>> >> Note: we resolved the domain name via DNS server. >>> >> >>> >> >>> >> For this we are using the command: >>> >> >>> >> ``` >>> >> >>> >> mount -t ceph [storagenode.storage.com]:6789:/ /backup -o >>> >> name=admin,secret=AQCM+8hjqzuZEhAAcuQc+onNKReq7MV+ykFirg== >>> >> >>> >> ``` >>> >> >>> >> >>> >> >>> >> We are getting the following logs in /var/log/messages: >>> >> >>> >> ``` >>> >> >>> >> Jan 24 17:23:17 localhost kernel: libceph: resolve ' >>> >> storagenode.storage.com' (ret=-3): failed >>> >> >>> >> Jan 24 17:23:17 localhost kernel: libceph: parse_ips bad ip ' >>> >> storagenode.storage.com:6789' >>> >> >>> >> ``` >>> >> >>> >> >>> >> >>> >> We also tried mounting ceph storage using IP of MON which is working >>> fine. >>> >> >>> >> >>> >> >>> >> Query: >>> >> >>> >> >>> >> Could you please help us out with how we can mount ceph using FQDN. >>> >> >>> >> >>> >> >>> >> My /etc/ceph/ceph.conf is as follows: >>> >> >>> >> [global] >>> >> >>> >> ms bind ipv6 = true >>> >> >>> >> ms bind ipv4 = false >>> >> >>> >> mon initial members = storagenode1,storagenode2,storagenode3 >>> >> >>> >> osd pool default crush rule = -1 >>> >> >>> >> fsid = 7969b8a3-1df7-4eae-8ccf-2e5794de87fe >>> >> >>> >> mon host = >>> >> >>> [v2:[abcd:abcd:abcd::21]:3300,v1:[abcd:abcd:abcd::21]:6789],[v2:[abcd:abcd:abcd::22]:3300,v1:[abcd:abcd:abcd::22]:6789],[v2:[abcd:abcd:abcd::23]:3300,v1:[abcd:abcd:abcd::23]:6789] >>> >> >>> >> public network = abcd:abcd:abcd::/64 >>> >> >>> >> cluster network = eff0:eff0:eff0::/64 >>> >> >>> >> >>> >> >>> >> [osd] >>> >> >>> >> osd memory target = 4294967296 >>> >> >>> >> >>> >> >>> >> [client.rgw.storagenode1.rgw0] >>> >> >>> >> host = storagenode1 >>> >> >>> >> keyring = /var/lib/ceph/radosgw/ceph-rgw.storagenode1.rgw0/keyring >>> >> >>> >> log file = /var/log/ceph/ceph-rgw-storagenode1.rgw0.log >>> >> >>> >> rgw frontends = beast endpoint=[abcd:abcd:abcd::21]:8080 >>> >> >>> >> rgw thread pool size = 512 >>> >> >>> >> -- >>> >> ~ Lokendra >>> >> skype: lokendrarathour >>> >> >>> >> >>> >> >>> > _______________________________________________ >>> > ceph-users mailing list -- ceph-users@xxxxxxx >>> > To unsubscribe send an email to ceph-users-leave@xxxxxxx >>> >>> _______________________________________________ >>> ceph-users mailing list -- ceph-users@xxxxxxx >>> To unsubscribe send an email to ceph-users-leave@xxxxxxx >>> >> >> >> -- >> ~ Lokendra >> skype: lokendrarathour >> >> >> > > -- > ~ Lokendra > skype: lokendrarathour > > > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx