Thank you for trying to reproduce my issue. I think I did that step: [centos@cnode-01 ~]$ ls -l /etc/ceph/ceph.client.admin.keyring -rw-------. 1 root root 151 May 10 23:20 /etc/ceph/ceph.client.admin.keyring [centos@cnode-01 ~]$ sudo cat /etc/ceph/ceph.client.admin.keyring [client.admin] key = AQC6v5lgi0JBAhAAJ9Duj11SufIKidydIgO82Q== caps mds = "allow *" caps mgr = "allow *" caps mon = "allow *" caps osd = "allow *" [centos@cnode-01 ~]$ sudo grep -r AQC6v5 /etc/ceph /tmp /etc/ceph/ceph.client.admin.keyring: key = AQC6v5lgi0JBAhAAJ9Duj11SufIKidydIgO82Q== /tmp/ceph.mon.keyring: key = AQC6v5lgi0JBAhAAJ9Duj11SufIKidydIgO82Q== [centos@cnode-01 ~]$ ________________________________ From: Eugen Block <eblock@xxxxxx> Sent: Thursday, May 13, 2021 12:37 AM To: Tuffli, Chuck <chuck.tuffli@xxxxxxx> Cc: ceph-users@xxxxxxx <ceph-users@xxxxxxx> Subject: Re: Re: monitor connection error I could reproduce your issue, it seems you're missing a keyring file in /etc/ceph. Did you miss step 9 from the manual deployment guide? sudo ceph-authtool --create-keyring /etc/ceph/ceph.client.admin.keyring --gen-key -n client.admin --cap mon 'allow *' --cap osd 'allow *' --cap mds 'allow *' --cap mgr 'allow *' Zitat von "Tuffli, Chuck" <chuck.tuffli@xxxxxxx>: >> -----Original Message----- >> From: Eugen Block [mailto:eblock@xxxxxx] >> Sent: Tuesday, May 11, 2021 11:39 PM >> To: ceph-users@xxxxxxx >> Subject: Re: monitor connection error >> >> Hi, >> >> > What is this error trying to tell me? TIA >> >> it tells you that the cluster is not reachable to the client, this >> can have various >> reasons. >> >> Can you show the output of your conf file? >> >> cat /etc/ceph/es-c1.conf > > [centos@cnode-01 ~]$ cat /etc/ceph/es-c1.conf > [global] > fsid = 3c5da069-2a03-4a5a-8396-53776286c858 > mon_initial_members = cnode-01,cnode-02,cnode-03 > mon_host = 192.168.122.39 > public_network = 192.168.122.0/24 > auth_cluster_required = cephx > auth_service_required = cephx > auth_client_required = cephx > osd_journal_size = 1024 > osd_pool_default_size = 3 > osd_pool_default_min_size = 2 > osd_pool_default_pg_num = 333 > osd_pool_default_pgp_num = 333 > osd_crush_chooseleaf_type = 1 > [centos@cnode-01 ~]$ > >> Is the monitor service up running? I take it you don't use cephadm >> yet so it's not >> a containerized environment? > > Correct, this is bare metal and not a containerized environment. And > I believe it is running: > [centos@cnode-01 ~]$ sudo systemctl --all | grep ceph > ceph-crash.service > loaded active running Ceph crash dump > collector > ceph-mon@cnode-01.service > loaded active running Ceph cluster > monitor daemon > system-ceph\x2dmon.slice > loaded active active > system-ceph\x2dmon.slice > ceph-mon.target > loaded active active ceph target > allowing to start/stop all ceph-mon@.service instances at once > ceph.target > loaded active active ceph target > allowing to start/stop all ceph*@.service instances at once > [centos@cnode-01 ~]$ > >> Regards, >> Eugen >> >> >> Zitat von "Tuffli, Chuck" <chuck.tuffli@xxxxxxx>: >> >> > Hi >> > >> > I'm new to ceph and have been following the Manual Deployment document >> > [1]. The process seems to work correctly until step 18 ("Verify that >> > the monitor is running"): >> > >> > [centos@cnode-01 ~]$ uname -a >> > Linux cnode-01 3.10.0-693.5.2.el7.x86_64 #1 SMP Fri Oct 20 20:32:50 >> > UTC 2017 x86_64 x86_64 x86_64 GNU/Linux >> > [centos@cnode-01 ~]$ ceph -v >> > ceph version 15.2.11 (e3523634d9c2227df9af89a4eac33d16738c49cb) >> > octopus (stable) >> > [centos@cnode-01 ~]$ sudo ceph --cluster es-c1 -s [errno 2] RADOS >> > object not found (error connecting to the cluster) >> > [centos@cnode-01 ~]$ >> > >> > What is this error trying to tell me? TIA >> > >> > [1] >> > INVALID URI REMOVED >> > nual-deployment/__;!!NpxR!1-v_Ql6E-l3P_E8DvIfk_YtknPrVFeZ5sFaPHLlsJVY8 >> > PmzP7kySRbr1rYqbFiZ1$ >> _______________________________________________ >> > ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an >> > email to ceph-users-leave@xxxxxxx >> >> >> _______________________________________________ >> ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send >> an email to >> ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx