---------- Forwarded message ----------
From: "hp cre" <hpcre1@xxxxxxxxx>
Date: 11 Apr 2016 15:50
Subject: Re: Ubuntu xenial and ceph jewel systemd
To: "James Page" <james.page@xxxxxxxxxx>
Cc:
2- run apt-get update && apt-get dist-upgrade (this step was not done on first trial)
4- dpkg -i ceph-deploy_1.5.31_all.deb
got errors of unmet dependencies, so i ran apt-get -f install. this installed all missing packages.
From: "hp cre" <hpcre1@xxxxxxxxx>
Date: 11 Apr 2016 15:50
Subject: Re: Ubuntu xenial and ceph jewel systemd
To: "James Page" <james.page@xxxxxxxxxx>
Cc:
Here is exactly what has been done (just started from scratch today):
1- install default xenial beta 22- run apt-get update && apt-get dist-upgrade (this step was not done on first trial)
after update, got warning as follows:
"W: plymouth: The plugin label.so is missing, the selected theme might not work as expected.
W: plymouth: You might want to install the plymouth-themes and plymouth-label package to fix this.
W: mdadm: /etc/mdadm/mdadm.conf defines no arrays.
"
"W: plymouth: The plugin label.so is missing, the selected theme might not work as expected.
W: plymouth: You might want to install the plymouth-themes and plymouth-label package to fix this.
W: mdadm: /etc/mdadm/mdadm.conf defines no arrays.
"
so i ran apt-get -y install plymouth-themes
3- wget http://download.ceph.com/debian-jewel/pool/main/c/ceph-deploy/ceph-deploy_1.5.31_all.deb4- dpkg -i ceph-deploy_1.5.31_all.deb
5- followed ceph docs preflight checklist (sudo file, ssh config file, ssh-copy-id, install ntp)
Followed the storage cluster quick start guide
6- ceph-deploy new xen1 (first node) --> all ok
6- ceph-deploy new xen1 (first node) --> all ok
7- edit ceph.conf --> osd pool default size = 2
8- ceph-deploy install --release=jewel xen1 --> all ok (this time it installed jewel 10.1.1, yesterday it was 10.1.0)
9- ceph-deploy mon create-initial --> same error:
wes@xen1:~/cl$ ceph-deploy mon create-initial
[ceph_deploy.conf][DEBUG ] found configuration file at: /home/wes/.cephdeploy.conf
[ceph_deploy.cli][INFO ] Invoked (1.5.31): /usr/bin/ceph-deploy mon create-initial
[ceph_deploy.cli][INFO ] ceph-deploy options:
[ceph_deploy.cli][INFO ] username : None
[ceph_deploy.cli][INFO ] verbose : False
[ceph_deploy.cli][INFO ] overwrite_conf : False
[ceph_deploy.cli][INFO ] subcommand : create-initial
[ceph_deploy.cli][INFO ] quiet : False
[ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0x7ffb88bfed88>
[ceph_deploy.cli][INFO ] cluster : ceph
[ceph_deploy.cli][INFO ] func : <function mon at 0x7ffb88bdcf50>
[ceph_deploy.cli][INFO ] ceph_conf : None
[ceph_deploy.cli][INFO ] keyrings : None
[ceph_deploy.cli][INFO ] default_release : False
[ceph_deploy.mon][DEBUG ] Deploying mon, cluster ceph hosts xen1
[ceph_deploy.mon][DEBUG ] detecting platform for host xen1 ...
[xen1][DEBUG ] connection detected need for sudo
[xen1][DEBUG ] connected to host: xen1
[xen1][DEBUG ] detect platform information from remote host
[xen1][DEBUG ] detect machine type
[xen1][DEBUG ] find the location of an executable
[ceph_deploy.mon][INFO ] distro info: Ubuntu 16.04 xenial
[xen1][DEBUG ] determining if provided host has same hostname in remote
[xen1][DEBUG ] get remote short hostname
[xen1][DEBUG ] deploying mon to xen1
[xen1][DEBUG ] get remote short hostname
[xen1][DEBUG ] remote hostname: xen1
[xen1][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[xen1][DEBUG ] create the mon path if it does not exist
[xen1][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-xen1/done
[xen1][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-xen1/done
[xen1][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-xen1.mon.keyring
[xen1][DEBUG ] create the monitor keyring file
[xen1][INFO ] Running command: sudo ceph-mon --cluster ceph --mkfs -i xen1 --keyring /var/lib/ceph/tmp/ceph-xen1.mon.keyring --setuser 64045 --setgroup 64045
[xen1][DEBUG ] ceph-mon: mon.noname-a 192.168.56.10:6789/0 is local, renaming to mon.xen1
[xen1][DEBUG ] ceph-mon: set fsid to d56c2ad9-dc66-4b6a-b269-e32eecc05571
[xen1][DEBUG ] ceph-mon: created monfs at /var/lib/ceph/mon/ceph-xen1 for mon.xen1
[xen1][INFO ] unlinking keyring file /var/lib/ceph/tmp/ceph-xen1.mon.keyring
[xen1][DEBUG ] create a done file to avoid re-doing the mon deployment
[xen1][DEBUG ] create the init path if it does not exist
[xen1][INFO ] Running command: sudo initctl emit ceph-mon cluster=ceph id=xen1
[xen1][ERROR ] Traceback (most recent call last):
[xen1][ERROR ] File "/usr/lib/python2.7/dist-packages/ceph_deploy/lib/vendor/remoto/process.py", line 119, in run
[xen1][ERROR ] reporting(conn, result, timeout)
[xen1][ERROR ] File "/usr/lib/python2.7/dist-packages/ceph_deploy/lib/vendor/remoto/log.py", line 13, in reporting
[xen1][ERROR ] received = result.receive(timeout)
[xen1][ERROR ] File "/usr/lib/python2.7/dist-packages/ceph_deploy/lib/vendor/remoto/lib/vendor/execnet/gateway_base.py", line 704, in receive
[xen1][ERROR ] raise self._getremoteerror() or EOFError()
[xen1][ERROR ] RemoteError: Traceback (most recent call last):
[xen1][ERROR ] File "/usr/lib/python2.7/dist-packages/ceph_deploy/lib/vendor/remoto/lib/vendor/execnet/gateway_base.py", line 1036, in executetask
[xen1][ERROR ] function(channel, **kwargs)
[xen1][ERROR ] File "<remote exec>", line 12, in _remote_run
[xen1][ERROR ] File "/usr/lib/python2.7/subprocess.py", line 711, in __init__
[xen1][ERROR ] errread, errwrite)
[xen1][ERROR ] File "/usr/lib/python2.7/subprocess.py", line 1340, in _execute_child
[xen1][ERROR ] raise child_exception
[xen1][ERROR ] OSError: [Errno 2] No such file or directory
[xen1][ERROR ]
[xen1][ERROR ]
[ceph_deploy.mon][ERROR ] Failed to execute command: initctl emit ceph-mon cluster=ceph id=xen1
[ceph_deploy][ERROR ] GenericError: Failed to create 1 monitors
9- ceph-deploy mon create-initial --> same error:
wes@xen1:~/cl$ ceph-deploy mon create-initial
[ceph_deploy.conf][DEBUG ] found configuration file at: /home/wes/.cephdeploy.conf
[ceph_deploy.cli][INFO ] Invoked (1.5.31): /usr/bin/ceph-deploy mon create-initial
[ceph_deploy.cli][INFO ] ceph-deploy options:
[ceph_deploy.cli][INFO ] username : None
[ceph_deploy.cli][INFO ] verbose : False
[ceph_deploy.cli][INFO ] overwrite_conf : False
[ceph_deploy.cli][INFO ] subcommand : create-initial
[ceph_deploy.cli][INFO ] quiet : False
[ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0x7ffb88bfed88>
[ceph_deploy.cli][INFO ] cluster : ceph
[ceph_deploy.cli][INFO ] func : <function mon at 0x7ffb88bdcf50>
[ceph_deploy.cli][INFO ] ceph_conf : None
[ceph_deploy.cli][INFO ] keyrings : None
[ceph_deploy.cli][INFO ] default_release : False
[ceph_deploy.mon][DEBUG ] Deploying mon, cluster ceph hosts xen1
[ceph_deploy.mon][DEBUG ] detecting platform for host xen1 ...
[xen1][DEBUG ] connection detected need for sudo
[xen1][DEBUG ] connected to host: xen1
[xen1][DEBUG ] detect platform information from remote host
[xen1][DEBUG ] detect machine type
[xen1][DEBUG ] find the location of an executable
[ceph_deploy.mon][INFO ] distro info: Ubuntu 16.04 xenial
[xen1][DEBUG ] determining if provided host has same hostname in remote
[xen1][DEBUG ] get remote short hostname
[xen1][DEBUG ] deploying mon to xen1
[xen1][DEBUG ] get remote short hostname
[xen1][DEBUG ] remote hostname: xen1
[xen1][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[xen1][DEBUG ] create the mon path if it does not exist
[xen1][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-xen1/done
[xen1][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-xen1/done
[xen1][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-xen1.mon.keyring
[xen1][DEBUG ] create the monitor keyring file
[xen1][INFO ] Running command: sudo ceph-mon --cluster ceph --mkfs -i xen1 --keyring /var/lib/ceph/tmp/ceph-xen1.mon.keyring --setuser 64045 --setgroup 64045
[xen1][DEBUG ] ceph-mon: mon.noname-a 192.168.56.10:6789/0 is local, renaming to mon.xen1
[xen1][DEBUG ] ceph-mon: set fsid to d56c2ad9-dc66-4b6a-b269-e32eecc05571
[xen1][DEBUG ] ceph-mon: created monfs at /var/lib/ceph/mon/ceph-xen1 for mon.xen1
[xen1][INFO ] unlinking keyring file /var/lib/ceph/tmp/ceph-xen1.mon.keyring
[xen1][DEBUG ] create a done file to avoid re-doing the mon deployment
[xen1][DEBUG ] create the init path if it does not exist
[xen1][INFO ] Running command: sudo initctl emit ceph-mon cluster=ceph id=xen1
[xen1][ERROR ] Traceback (most recent call last):
[xen1][ERROR ] File "/usr/lib/python2.7/dist-packages/ceph_deploy/lib/vendor/remoto/process.py", line 119, in run
[xen1][ERROR ] reporting(conn, result, timeout)
[xen1][ERROR ] File "/usr/lib/python2.7/dist-packages/ceph_deploy/lib/vendor/remoto/log.py", line 13, in reporting
[xen1][ERROR ] received = result.receive(timeout)
[xen1][ERROR ] File "/usr/lib/python2.7/dist-packages/ceph_deploy/lib/vendor/remoto/lib/vendor/execnet/gateway_base.py", line 704, in receive
[xen1][ERROR ] raise self._getremoteerror() or EOFError()
[xen1][ERROR ] RemoteError: Traceback (most recent call last):
[xen1][ERROR ] File "/usr/lib/python2.7/dist-packages/ceph_deploy/lib/vendor/remoto/lib/vendor/execnet/gateway_base.py", line 1036, in executetask
[xen1][ERROR ] function(channel, **kwargs)
[xen1][ERROR ] File "<remote exec>", line 12, in _remote_run
[xen1][ERROR ] File "/usr/lib/python2.7/subprocess.py", line 711, in __init__
[xen1][ERROR ] errread, errwrite)
[xen1][ERROR ] File "/usr/lib/python2.7/subprocess.py", line 1340, in _execute_child
[xen1][ERROR ] raise child_exception
[xen1][ERROR ] OSError: [Errno 2] No such file or directory
[xen1][ERROR ]
[xen1][ERROR ]
[ceph_deploy.mon][ERROR ] Failed to execute command: initctl emit ceph-mon cluster=ceph id=xen1
[ceph_deploy][ERROR ] GenericError: Failed to create 1 monitors
On 11 April 2016 at 14:05, hp cre <hpcre1@xxxxxxxxx> wrote:
In the process of reproducing it now. I'll attach a full command log
On 11 Apr 2016 11:42, "James Page" <james.page@xxxxxxxxxx> wrote:It would be handy to get visibility of your deployment log data; I'm not currently able to reproduce your issue deploying ceph using ceph-deploy on a small three node install running xenial - its correctly detecting systemd and using systemctl instead of initctl.On Mon, 11 Apr 2016 at 10:18 James Page <james.page@xxxxxxxxxx> wrote:On Mon, 11 Apr 2016 at 10:02 hp cre <hpcre1@xxxxxxxxx> wrote:Hello James,
It's a default install of xenial server beta 2 release. Created a user then followed the ceph installation quick start exactly as it is.
Ceph-deploy version 1.5.31 was used as follows
1- ceph-deploy new node1
2- ceph-deploy install --release jewel node1
3- ceph-deploy mon create-initialStep 3 gave error in Python scripts. Meaning it could not find initctl command. Searched for this command and found out our belongs to upstart.
I suspect that ceph-deploy is not playing nicely with systemd based Ubuntu releases - I'll take a look now..
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com