Ceph version Quincy. But now I am able to resolve the issue. During mount i will not pass any monitor details, it will be auto-discovered via SRV. On Tue, Apr 11, 2023 at 6:09 PM Eugen Block <eblock@xxxxxx> wrote: > What ceph version is this? Could it be this bug [1]? Although the > error message is different, not sure if it could be the same issue, > and I don't have anything to test ipv6 with. > > [1] https://tracker.ceph.com/issues/47300 > > Zitat von Lokendra Rathour <lokendrarathour@xxxxxxxxx>: > > > Hi All, > > Requesting any inputs around the issue raised. > > > > Best Regards, > > Lokendra > > > > On Tue, 24 Jan, 2023, 7:32 pm Lokendra Rathour, < > lokendrarathour@xxxxxxxxx> > > wrote: > > > >> Hi Team, > >> > >> > >> > >> We have a ceph cluster with 3 storage nodes: > >> > >> 1. storagenode1 - abcd:abcd:abcd::21 > >> > >> 2. storagenode2 - abcd:abcd:abcd::22 > >> > >> 3. storagenode3 - abcd:abcd:abcd::23 > >> > >> > >> > >> The requirement is to mount ceph using the domain name of MON node: > >> > >> Note: we resolved the domain name via DNS server. > >> > >> > >> For this we are using the command: > >> > >> ``` > >> > >> mount -t ceph [storagenode.storage.com]:6789:/ /backup -o > >> name=admin,secret=AQCM+8hjqzuZEhAAcuQc+onNKReq7MV+ykFirg== > >> > >> ``` > >> > >> > >> > >> We are getting the following logs in /var/log/messages: > >> > >> ``` > >> > >> Jan 24 17:23:17 localhost kernel: libceph: resolve ' > >> storagenode.storage.com' (ret=-3): failed > >> > >> Jan 24 17:23:17 localhost kernel: libceph: parse_ips bad ip ' > >> storagenode.storage.com:6789' > >> > >> ``` > >> > >> > >> > >> We also tried mounting ceph storage using IP of MON which is working > fine. > >> > >> > >> > >> Query: > >> > >> > >> Could you please help us out with how we can mount ceph using FQDN. > >> > >> > >> > >> My /etc/ceph/ceph.conf is as follows: > >> > >> [global] > >> > >> ms bind ipv6 = true > >> > >> ms bind ipv4 = false > >> > >> mon initial members = storagenode1,storagenode2,storagenode3 > >> > >> osd pool default crush rule = -1 > >> > >> fsid = 7969b8a3-1df7-4eae-8ccf-2e5794de87fe > >> > >> mon host = > >> > [v2:[abcd:abcd:abcd::21]:3300,v1:[abcd:abcd:abcd::21]:6789],[v2:[abcd:abcd:abcd::22]:3300,v1:[abcd:abcd:abcd::22]:6789],[v2:[abcd:abcd:abcd::23]:3300,v1:[abcd:abcd:abcd::23]:6789] > >> > >> public network = abcd:abcd:abcd::/64 > >> > >> cluster network = eff0:eff0:eff0::/64 > >> > >> > >> > >> [osd] > >> > >> osd memory target = 4294967296 > >> > >> > >> > >> [client.rgw.storagenode1.rgw0] > >> > >> host = storagenode1 > >> > >> keyring = /var/lib/ceph/radosgw/ceph-rgw.storagenode1.rgw0/keyring > >> > >> log file = /var/log/ceph/ceph-rgw-storagenode1.rgw0.log > >> > >> rgw frontends = beast endpoint=[abcd:abcd:abcd::21]:8080 > >> > >> rgw thread pool size = 512 > >> > >> -- > >> ~ Lokendra > >> skype: lokendrarathour > >> > >> > >> > > _______________________________________________ > > ceph-users mailing list -- ceph-users@xxxxxxx > > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > -- ~ Lokendra skype: lokendrarathour _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx