Re: Cephadm Upgrade from Octopus to Pacific

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Yes that works for me, looping through the nodes and doing a is a quick
work around.

ssh $node mkdir /usr/lib/sysctl.d

does allow the upgrade to complete. As I've probably said before I'm
using Debian Buster as CentOS 7 was not 100% happy even with Octopus
(although it just about worked) and CentOs 8 does not support the hardware
I've got (the disks are not detected, and I can't find the right
drivers....)

I suspect I've not got to do some tidying up before I continue but this
does look smoother than when I tried with 16.2.0, which was 4 months ago.

Thanks

Peter.

On Fri, 6 Aug 2021 at 14:40, Dimitri Savineau <dsavinea@xxxxxxxxxx> wrote:

> Looks related to https://tracker.ceph.com/issues/51620
>
> Hopefully this will be backported to Pacific and included in 16.2.6
>
> Regards,
>
> Dimitri
>
> On Fri, Aug 6, 2021 at 9:21 AM Arnaud MARTEL <
> arnaud.martel@xxxxxxxxxxxxxxxxxxxx> wrote:
>
> > Peter,
> >
> > I had the same error and my workaround was to manually create
> > /usr/lib/sysctl.d directory on all nodes, then resume the upgrade
> >
> > Arnaud Martel
> > ----- Mail original -----
> > De: "Peter Childs" <pchilds@xxxxxxx>
> > À: "ceph-users" <ceph-users@xxxxxxx>
> > Envoyé: Vendredi 6 Août 2021 15:03:20
> > Objet:  Cephadm Upgrade from Octopus to Pacific
> >
> > I'm attempting to upgrade my large cephadm deployed cluster with 1600 osd
> > from octopus to pacific.
> >
> > Given this did not work very well when I first tried I decided to break
> off
> > 3 nodes and create a small "test" cluster to see how badly it would fail.
> >
> > This is upgrading from 15.2.13 to 16.2.5 with a small 3 node cluster.
> with
> > only 33 osd's (not 42 nodes and 1600 osds)
> >
> > using
> >
> > ceph orch upgrade start --ceph-version 16.2.5
> >
> > so far I've got as far as the first osd failing the mgrs and mons
> upgraded
> > quite quickly, but the very first osd failed.
> >
> > 8/6/21 1:51:19 PM[ERR]Upgrade: Paused due to UPGRADE_REDEPLOY_DAEMON:
> > Upgrading daemon osd.3 on host drywood10 failed.
> >
> > 8/6/21 1:51:19 PM[ERR]cephadm exited with an error code: 1,
> stderr:Redeploy
> > daemon osd.3 ... Traceback (most recent call last): File
> >
> >
> "/var/lib/ceph/43fd7d2e-f693-11eb-990a-a4bf01112a34/cephadm.d4237e4639c108308fe13147b1c08af93c3d5724d9ff21ae797eb4b78fea3931",
> > line 8230, in <module> main() File
> >
> >
> "/var/lib/ceph/43fd7d2e-f693-11eb-990a-a4bf01112a34/cephadm.d4237e4639c108308fe13147b1c08af93c3d5724d9ff21ae797eb4b78fea3931",
> > line 8218, in main r = ctx.func(ctx) File
> >
> >
> "/var/lib/ceph/43fd7d2e-f693-11eb-990a-a4bf01112a34/cephadm.d4237e4639c108308fe13147b1c08af93c3d5724d9ff21ae797eb4b78fea3931",
> > line 1759, in _default_image return func(ctx) File
> >
> >
> "/var/lib/ceph/43fd7d2e-f693-11eb-990a-a4bf01112a34/cephadm.d4237e4639c108308fe13147b1c08af93c3d5724d9ff21ae797eb4b78fea3931",
> > line 4326, in command_deploy ports=daemon_ports) File
> >
> >
> "/var/lib/ceph/43fd7d2e-f693-11eb-990a-a4bf01112a34/cephadm.d4237e4639c108308fe13147b1c08af93c3d5724d9ff21ae797eb4b78fea3931",
> > line 2632, in deploy_daemon c, osd_fsid=osd_fsid, ports=ports) File
> >
> >
> "/var/lib/ceph/43fd7d2e-f693-11eb-990a-a4bf01112a34/cephadm.d4237e4639c108308fe13147b1c08af93c3d5724d9ff21ae797eb4b78fea3931",
> > line 2801, in deploy_daemon_units install_sysctl(ctx, fsid, daemon_type)
> > File
> >
> >
> "/var/lib/ceph/43fd7d2e-f693-11eb-990a-a4bf01112a34/cephadm.d4237e4639c108308fe13147b1c08af93c3d5724d9ff21ae797eb4b78fea3931",
> > line 2963, in install_sysctl _write(conf, lines) File
> >
> >
> "/var/lib/ceph/43fd7d2e-f693-11eb-990a-a4bf01112a34/cephadm.d4237e4639c108308fe13147b1c08af93c3d5724d9ff21ae797eb4b78fea3931",
> > line 2948, in _write with open(conf, 'w') as f: FileNotFoundError: [Errno
> > 2] No such file or directory:
> > '/usr/lib/sysctl.d/90-ceph-43fd7d2e-f693-11eb-990a-a4bf01112a34-osd.conf'
> > Traceback (most recent call last): File
> > "/usr/share/ceph/mgr/cephadm/serve.py", line 1347, in _remote_connection
> > yield (conn, connr) File "/usr/share/ceph/mgr/cephadm/serve.py", line
> 1244,
> > in _run_cephadm code, '\n'.join(err)))
> > orchestrator._interface.OrchestratorError: cephadm exited with an error
> > code: 1, stderr:Redeploy daemon osd.3 ... Traceback (most recent call
> > last): File
> >
> >
> "/var/lib/ceph/43fd7d2e-f693-11eb-990a-a4bf01112a34/cephadm.d4237e4639c108308fe13147b1c08af93c3d5724d9ff21ae797eb4b78fea3931",
> > line 8230, in <module> main() File
> >
> >
> "/var/lib/ceph/43fd7d2e-f693-11eb-990a-a4bf01112a34/cephadm.d4237e4639c108308fe13147b1c08af93c3d5724d9ff21ae797eb4b78fea3931",
> > line 8218, in main r = ctx.func(ctx) File
> >
> >
> "/var/lib/ceph/43fd7d2e-f693-11eb-990a-a4bf01112a34/cephadm.d4237e4639c108308fe13147b1c08af93c3d5724d9ff21ae797eb4b78fea3931",
> > line 1759, in _default_image return func(ctx) File
> >
> >
> "/var/lib/ceph/43fd7d2e-f693-11eb-990a-a4bf01112a34/cephadm.d4237e4639c108308fe13147b1c08af93c3d5724d9ff21ae797eb4b78fea3931",
> > line 4326, in command_deploy ports=daemon_ports) File
> >
> >
> "/var/lib/ceph/43fd7d2e-f693-11eb-990a-a4bf01112a34/cephadm.d4237e4639c108308fe13147b1c08af93c3d5724d9ff21ae797eb4b78fea3931",
> > line 2632, in deploy_daemon c, osd_fsid=osd_fsid, ports=ports) File
> >
> >
> "/var/lib/ceph/43fd7d2e-f693-11eb-990a-a4bf01112a34/cephadm.d4237e4639c108308fe13147b1c08af93c3d5724d9ff21ae797eb4b78fea3931",
> > line 2801, in deploy_daemon_units install_sysctl(ctx, fsid, daemon_type)
> > File
> >
> >
> "/var/lib/ceph/43fd7d2e-f693-11eb-990a-a4bf01112a34/cephadm.d4237e4639c108308fe13147b1c08af93c3d5724d9ff21ae797eb4b78fea3931",
> > line 2963, in install_sysctl _write(conf, lines) File
> >
> >
> "/var/lib/ceph/43fd7d2e-f693-11eb-990a-a4bf01112a34/cephadm.d4237e4639c108308fe13147b1c08af93c3d5724d9ff21ae797eb4b78fea3931",
> > line 2948, in _write with open(conf, 'w') as f: FileNotFoundError: [Errno
> > 2] No such file or directory:
> > '/usr/lib/sysctl.d/90-ceph-43fd7d2e-f693-11eb-990a-a4bf01112a34-osd.conf'
> >
> >
> > The good news this is a pre-production proof of concept cluster still so
> > I'm attempting to iron out issues, before we try and make it a production
> > service.
> >
> > Any ideas would be helpful.
> >
> > I guess deploy might be an option but that does not feel very future
> proof.
> >
> >
> > Thanks
> >
> >
> > Peter Childs
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@xxxxxxx
> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@xxxxxxx
> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
> >
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux