Hi Team, The mount at the client side should be independent of Ceph, but here in this case of DNS SRV-based mount, we see that the Ceph common utility is needed. What can be the reason for the same, any inputs in this direction would be helpful. Best Regards, Lokendra On Sun, Apr 16, 2023 at 10:11 AM Lokendra Rathour <lokendrarathour@xxxxxxxxx> wrote: > Hi . > Any input will be of great help. > Thanks once again. > Lokendra > > On Fri, 14 Apr, 2023, 3:47 pm Lokendra Rathour, <lokendrarathour@xxxxxxxxx> > wrote: > >> Hi Team, >> their is one additional observation. >> Mount as the client is working fine from one of the Ceph nodes. >> Command *: sudo mount -t ceph :/ /mnt/imgs -o >> name=foo,secret=AQABDzRkTaJCEhAAC7rC6E68ofwdfULnx6qX/VDA== * >> >> *we are not passing the Monitor address, instead, DNS SRV is configured >> as per:* >> https://docs.ceph.com/en/quincy/rados/configuration/mon-lookup-dns/ >> >> mount works fine in this case. >> >> #### >> >> But if we try to mount from the other Location i.e from another >> VM/client(non-Ceph Node) >> we are getting the error : >> mount -t ceph :/ /mnt/imgs -o >> name=foo,secret=AQABDzRkTaJCEhAAC7rC6E68ofwULnx6qX/VDA== -v >> *mount: /mnt/image: mount point does not exist.* >> >> the document says that if we do not pass the monitor address, it tries >> discovering the monitor address from DNS Servers, but in actual it is not >> happening. >> >> >> >> On Tue, Apr 11, 2023 at 6:48 PM Lokendra Rathour < >> lokendrarathour@xxxxxxxxx> wrote: >> >>> Ceph version Quincy. >>> >>> But now I am able to resolve the issue. >>> >>> During mount i will not pass any monitor details, it will be >>> auto-discovered via SRV. >>> >>> On Tue, Apr 11, 2023 at 6:09 PM Eugen Block <eblock@xxxxxx> wrote: >>> >>>> What ceph version is this? Could it be this bug [1]? Although the >>>> error message is different, not sure if it could be the same issue, >>>> and I don't have anything to test ipv6 with. >>>> >>>> [1] https://tracker.ceph.com/issues/47300 >>>> >>>> Zitat von Lokendra Rathour <lokendrarathour@xxxxxxxxx>: >>>> >>>> > Hi All, >>>> > Requesting any inputs around the issue raised. >>>> > >>>> > Best Regards, >>>> > Lokendra >>>> > >>>> > On Tue, 24 Jan, 2023, 7:32 pm Lokendra Rathour, < >>>> lokendrarathour@xxxxxxxxx> >>>> > wrote: >>>> > >>>> >> Hi Team, >>>> >> >>>> >> >>>> >> >>>> >> We have a ceph cluster with 3 storage nodes: >>>> >> >>>> >> 1. storagenode1 - abcd:abcd:abcd::21 >>>> >> >>>> >> 2. storagenode2 - abcd:abcd:abcd::22 >>>> >> >>>> >> 3. storagenode3 - abcd:abcd:abcd::23 >>>> >> >>>> >> >>>> >> >>>> >> The requirement is to mount ceph using the domain name of MON node: >>>> >> >>>> >> Note: we resolved the domain name via DNS server. >>>> >> >>>> >> >>>> >> For this we are using the command: >>>> >> >>>> >> ``` >>>> >> >>>> >> mount -t ceph [storagenode.storage.com]:6789:/ /backup -o >>>> >> name=admin,secret=AQCM+8hjqzuZEhAAcuQc+onNKReq7MV+ykFirg== >>>> >> >>>> >> ``` >>>> >> >>>> >> >>>> >> >>>> >> We are getting the following logs in /var/log/messages: >>>> >> >>>> >> ``` >>>> >> >>>> >> Jan 24 17:23:17 localhost kernel: libceph: resolve ' >>>> >> storagenode.storage.com' (ret=-3): failed >>>> >> >>>> >> Jan 24 17:23:17 localhost kernel: libceph: parse_ips bad ip ' >>>> >> storagenode.storage.com:6789' >>>> >> >>>> >> ``` >>>> >> >>>> >> >>>> >> >>>> >> We also tried mounting ceph storage using IP of MON which is working >>>> fine. >>>> >> >>>> >> >>>> >> >>>> >> Query: >>>> >> >>>> >> >>>> >> Could you please help us out with how we can mount ceph using FQDN. >>>> >> >>>> >> >>>> >> >>>> >> My /etc/ceph/ceph.conf is as follows: >>>> >> >>>> >> [global] >>>> >> >>>> >> ms bind ipv6 = true >>>> >> >>>> >> ms bind ipv4 = false >>>> >> >>>> >> mon initial members = storagenode1,storagenode2,storagenode3 >>>> >> >>>> >> osd pool default crush rule = -1 >>>> >> >>>> >> fsid = 7969b8a3-1df7-4eae-8ccf-2e5794de87fe >>>> >> >>>> >> mon host = >>>> >> >>>> [v2:[abcd:abcd:abcd::21]:3300,v1:[abcd:abcd:abcd::21]:6789],[v2:[abcd:abcd:abcd::22]:3300,v1:[abcd:abcd:abcd::22]:6789],[v2:[abcd:abcd:abcd::23]:3300,v1:[abcd:abcd:abcd::23]:6789] >>>> >> >>>> >> public network = abcd:abcd:abcd::/64 >>>> >> >>>> >> cluster network = eff0:eff0:eff0::/64 >>>> >> >>>> >> >>>> >> >>>> >> [osd] >>>> >> >>>> >> osd memory target = 4294967296 >>>> >> >>>> >> >>>> >> >>>> >> [client.rgw.storagenode1.rgw0] >>>> >> >>>> >> host = storagenode1 >>>> >> >>>> >> keyring = /var/lib/ceph/radosgw/ceph-rgw.storagenode1.rgw0/keyring >>>> >> >>>> >> log file = /var/log/ceph/ceph-rgw-storagenode1.rgw0.log >>>> >> >>>> >> rgw frontends = beast endpoint=[abcd:abcd:abcd::21]:8080 >>>> >> >>>> >> rgw thread pool size = 512 >>>> >> >>>> >> -- >>>> >> ~ Lokendra >>>> >> skype: lokendrarathour >>>> >> >>>> >> >>>> >> >>>> > _______________________________________________ >>>> > ceph-users mailing list -- ceph-users@xxxxxxx >>>> > To unsubscribe send an email to ceph-users-leave@xxxxxxx >>>> >>>> _______________________________________________ >>>> ceph-users mailing list -- ceph-users@xxxxxxx >>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx >>>> >>> >>> >>> -- >>> ~ Lokendra >>> skype: lokendrarathour >>> >>> >>> >> >> -- >> ~ Lokendra >> skype: lokendrarathour >> >> >> -- ~ Lokendra skype: lokendrarathour _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx