Hi, looks like you haven’t run the ceph-deploy command with the same user name and may be not the same current working directory. This could explain your problem. Make sure the other daemons have a mgr cap authorisation. You can find on this ML details about MGR caps being incorrect for OSDs and MONs after a Jewel to Luminous upgrade. The output of a ceph auth list command should help you find out if it’s the case. Are your ceph daemons still running? What does a ceph daemon mon.$(hostname -s) quorum_status gives you from a MON server. JC > On Feb 28, 2018, at 10:05, Georgios Dimitrakakis <giorgis@xxxxxxxxxxxx> wrote: > > > Indeed John, > > you are right! I have updated "ceph-deploy" (which was installed via "pip" that's why wasn't updated with the rest ceph packages) but now it complaints that keys are missing > > $ ceph-deploy mgr create controller > [ceph_deploy.conf][DEBUG ] found configuration file at: /home/user/.cephdeploy.conf > [ceph_deploy.cli][INFO ] Invoked (2.0.0): /usr/bin/ceph-deploy mgr create controller > [ceph_deploy.cli][INFO ] ceph-deploy options: > [ceph_deploy.cli][INFO ] username : None > [ceph_deploy.cli][INFO ] verbose : False > [ceph_deploy.cli][INFO ] mgr : [('controller', 'controller')] > [ceph_deploy.cli][INFO ] overwrite_conf : False > [ceph_deploy.cli][INFO ] subcommand : create > [ceph_deploy.cli][INFO ] quiet : False > [ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0x1d42bd8> > [ceph_deploy.cli][INFO ] cluster : ceph > [ceph_deploy.cli][INFO ] func : <function mgr at 0x1cce500> > [ceph_deploy.cli][INFO ] ceph_conf : None > [ceph_deploy.cli][INFO ] default_release : False > [ceph_deploy.mgr][DEBUG ] Deploying mgr, cluster ceph hosts controller:controller > [ceph_deploy][ERROR ] RuntimeError: bootstrap-mgr keyring not found; run 'gatherkeys' > > > and I cannot get the keys... > > > > $ ceph-deploy gatherkeys controller > [ceph_deploy.conf][DEBUG ] found configuration file at: /home/user/.cephdeploy.conf > [ceph_deploy.cli][INFO ] Invoked (2.0.0): /usr/bin/ceph-deploy gatherkeys controller > [ceph_deploy.cli][INFO ] ceph-deploy options: > [ceph_deploy.cli][INFO ] username : None > [ceph_deploy.cli][INFO ] verbose : False > [ceph_deploy.cli][INFO ] overwrite_conf : False > [ceph_deploy.cli][INFO ] quiet : False > [ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0x199f290> > [ceph_deploy.cli][INFO ] cluster : ceph > [ceph_deploy.cli][INFO ] mon : ['controller'] > [ceph_deploy.cli][INFO ] func : <function gatherkeys at 0x198b2a8> > [ceph_deploy.cli][INFO ] ceph_conf : None > [ceph_deploy.cli][INFO ] default_release : False > [ceph_deploy.gatherkeys][INFO ] Storing keys in temp directory /tmp/tmpPQ895t > [controller][DEBUG ] connection detected need for sudo > [controller][DEBUG ] connected to host: controller > [controller][DEBUG ] detect platform information from remote host > [controller][DEBUG ] detect machine type > [controller][DEBUG ] get remote short hostname > [controller][DEBUG ] fetch remote file > [ceph_deploy.gatherkeys][WARNIN] No mon key found in host: controller > [ceph_deploy.gatherkeys][ERROR ] Failed to connect to host:controller > [ceph_deploy.gatherkeys][INFO ] Destroy temp directory /tmp/tmpPQ895t > [ceph_deploy][ERROR ] RuntimeError: Failed to connect any mon > > > > >> On Wed, Feb 28, 2018 at 5:21 PM, Georgios Dimitrakakis >> <giorgis@xxxxxxxxxxxx> wrote: >>> All, >>> >>> I have updated my test ceph cluster from Jewer (10.2.10) to Luminous >>> (12.2.4) using CentOS packages. >>> >>> I have updated all packages, restarted all services with the proper order >>> but I get a warning that the Manager Daemon doesn't exist. >>> >>> Here is the output: >>> >>> # ceph -s >>> cluster: >>> id: d357a551-5b7a-4501-8d8f-009c63b2c972 >>> health: HEALTH_WARN >>> no active mgr >>> >>> services: >>> mon: 1 daemons, quorum controller >>> mgr: no daemons active >>> osd: 2 osds: 2 up, 2 in >>> >>> data: >>> pools: 0 pools, 0 pgs >>> objects: 0 objects, 0 bytes >>> usage: 0 kB used, 0 kB / 0 kB avail >>> pgs: >>> >>> >>> While at the same time the system service is up and running >>> >>> # systemctl status ceph-mgr.target >>> ● ceph-mgr.target - ceph target allowing to start/stop all ceph-mgr@.service >>> instances at once >>> Loaded: loaded (/usr/lib/systemd/system/ceph-mgr.target; enabled; vendor >>> preset: enabled) >>> Active: active since Wed 2018-02-28 18:57:13 EET; 12min ago >>> >>> >>> I understand that I have to add a new MGR but when I try to do it via >>> "ceph-deploy" it fails with the following error: >>> >>> >>> # ceph-deploy mgr create controller >>> usage: ceph-deploy [-h] [-v | -q] [--version] [--username USERNAME] >>> [--overwrite-conf] [--cluster NAME] [--ceph-conf >>> CEPH_CONF] >>> COMMAND ... >>> ceph-deploy: error: argument COMMAND: invalid choice: 'mgr' (choose from >>> 'new', 'install', 'rgw', 'mon', 'mds', 'gatherkeys', 'disk', 'osd', 'admin', >>> 'repo', 'config', 'uninstall', 'purge', 'purgedata', 'calamari', >>> 'forgetkeys', 'pkg') >> >> You probably have an older version of ceph-deploy, from before it knew >> how to create mgr daemons. >> >> John >> >>> >>> >>> where "controller" is the node where ceph monitor is already running. >>> >>> >>> Any ideas why I cannot do it via "ceph-deploy" and what do I have to do to >>> have it back in a healthy state? >>> >>> >>> I am running CentOS 7.4.1708 (Core). >>> >>> Best, >>> >>> G. >>> >>> _______________________________________________ >>> ceph-users mailing list >>> ceph-users@xxxxxxxxxxxxxx >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > -- > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com