Hi Lokendra, To make monitor looked up through dns, ceph-mon also need to be resolved correctly by dns server just like _ceph-mon._tcp. And ceph-mon is default service name, which doesn’t need to be in the conf file anyway. Ben > 2023年2月3日 12:14,Lokendra Rathour <lokendrarathour@xxxxxxxxx> 写道: > > Hi Robert and Team, > > > > Thank you for the help. We had previously referred to the link: > https://docs.ceph.com/en/quincy/rados/configuration/mon-lookup-dns/ > But we were not able to configure mon_dns_srv_name correctly. > > > > We find the following link: > https://access.redhat.com/documentation/en-us/red_hat_ceph_storage/4/html/configuration_guide/ceph-monitor-configuration > > > > Which gives just a little more information about the DNS lookup. > > > > After following the link, we updated the ceph.conf as follows: > ``` > [root@storagenode3 ~]# cat /etc/ceph/ceph.conf > [global] > ms bind ipv6 = true > ms bind ipv4 = false > mon initial members = storagenode1,storagenode2,storagenode3 > osd pool default crush rule = -1 > mon dns srv name = ceph-mon > fsid = ce479912-a277-45b6-87b1-203d3e43d776 > public network = abcd:abcd:abcd::/64 > cluster network = eff0:eff0:eff0::/64 > > > > [osd] > osd memory target = 4294967296 > > > > [client.rgw.storagenode3.rgw0] > host = storagenode3 > keyring = /var/lib/ceph/radosgw/ceph-rgw.storagenode3.rgw0/keyring > log file = /var/log/ceph/ceph-rgw-storagenode3.rgw0.log > rgw frontends = beast endpoint=[abcd:abcd:abcd::23]:8080 > rgw thread pool size = 512 > > > > [root@storagenode3 ~]# > ``` > > We also updated the dns server as follows: > > ``` > storagenode1.storage.com IN AAAA abcd:abcd:abcd::21 > storagenode2.storage.com IN AAAA abcd:abcd:abcd::22 > storagenode3.storage.com IN AAAA abcd:abcd:abcd::23 > > > > _ceph-mon._tcp.storage.com 60 IN SRV 10 60 6789 storagenode1.storage.com > _ceph-mon._tcp.storage.com 60 IN SRV 10 60 6789 storagenode2.storage.com > _ceph-mon._tcp.storage.com 60 IN SRV 10 60 6789 storagenode3.storage.com > _ceph-mon._tcp.storage.com 60 IN SRV 10 60 3300 storagenode1.storage.com > _ceph-mon._tcp.storage.com 60 IN SRV 10 60 3300 storagenode2.storage.com > _ceph-mon._tcp.storage.com 60 IN SRV 10 60 3300 storagenode3.storage.com > > > ``` > > But when we run the command ceph -s, we get the following error: > > ``` > [root@storagenode3 ~]# ceph -s > unable to get monitor info from DNS SRV with service name: ceph-mon > 2023-02-02T15:18:14.098+0530 7f1313a58700 -1 failed for service > _ceph-mon._tcp > 2023-02-02T15:18:14.098+0530 7f1313a58700 -1 monclient: > get_monmap_and_config cannot identify monitors to contact > [errno 2] RADOS object not found (error connecting to the cluster) > [root@storagenode3 ~]# > ``` > > Could you please help us to configure the server using mon_dns_srv_name > correctly? > > > > On Wed, Jan 25, 2023 at 9:06 PM John Mulligan <phlogistonjohn@xxxxxxxxxxxxx> > wrote: > >> On Tuesday, January 24, 2023 9:02:41 AM EST Lokendra Rathour wrote: >>> Hi Team, >>> >>> >>> >>> We have a ceph cluster with 3 storage nodes: >>> >>> 1. storagenode1 - abcd:abcd:abcd::21 >>> >>> 2. storagenode2 - abcd:abcd:abcd::22 >>> >>> 3. storagenode3 - abcd:abcd:abcd::23 >>> >>> >>> >>> The requirement is to mount ceph using the domain name of MON node: >>> >>> Note: we resolved the domain name via DNS server. >>> >>> >>> For this we are using the command: >>> >>> ``` >>> >>> mount -t ceph [storagenode.storage.com]:6789:/ /backup -o >>> name=admin,secret=AQCM+8hjqzuZEhAAcuQc+onNKReq7MV+ykFirg== >>> >>> ``` >>> >>> >>> >>> We are getting the following logs in /var/log/messages: >>> >>> ``` >>> >>> Jan 24 17:23:17 localhost kernel: libceph: resolve ' >> storagenode.storage.com' >>> (ret=-3): failed >>> >>> Jan 24 17:23:17 localhost kernel: libceph: parse_ips bad ip ' >>> storagenode.storage.com:6789' >>> >>> ``` >>> >> >> >> I saw a similar log message recently when I had forgotten to install the >> ceph >> mount helper. >> Check to see if you have a binary 'mount.ceph' on the system. If you don't >> try >> to install it from packages. On fedora I needed to install 'ceph-common'. >> >> >>> >>> >>> We also tried mounting ceph storage using IP of MON which is working >> fine. >>> >>> >>> >>> Query: >>> >>> >>> Could you please help us out with how we can mount ceph using FQDN. >>> >>> >>> >>> My /etc/ceph/ceph.conf is as follows: >>> >>> [global] >>> >>> ms bind ipv6 = true >>> >>> ms bind ipv4 = false >>> >>> mon initial members = storagenode1,storagenode2,storagenode3 >>> >>> osd pool default crush rule = -1 >>> >>> fsid = 7969b8a3-1df7-4eae-8ccf-2e5794de87fe >>> >>> mon host = >>> >> [v2:[abcd:abcd:abcd::21]:3300,v1:[abcd:abcd:abcd::21]:6789],[v2:[abcd:abcd:a >>> >> bcd::22]:3300,v1:[abcd:abcd:abcd::22]:6789],[v2:[abcd:abcd:abcd::23]:3300,v1 >>> :[abcd:abcd:abcd::23]:6789] >>> >>> public network = abcd:abcd:abcd::/64 >>> >>> cluster network = eff0:eff0:eff0::/64 >>> >>> >>> >>> [osd] >>> >>> osd memory target = 4294967296 >>> >>> >>> >>> [client.rgw.storagenode1.rgw0] >>> >>> host = storagenode1 >>> >>> keyring = /var/lib/ceph/radosgw/ceph-rgw.storagenode1.rgw0/keyring >>> >>> log file = /var/log/ceph/ceph-rgw-storagenode1.rgw0.log >>> >>> rgw frontends = beast endpoint=[abcd:abcd:abcd::21]:8080 >>> >>> rgw thread pool size = 512 >> >> >> >> >> > > -- > ~ Lokendra > skype: lokendrarathour > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx