Re: v15.2.0 Octopus released

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 2020-04-02 16:40, konstantin.ilyasov@xxxxxxxxxxxxxx wrote:
> I have done it.
> I am not sure, if i didn’t miss something, but i upgraded test cluster from CentOs7.7.1908+Ceph14.2.8 to Debian10.3+Ceph15.2.0.
> 
> Preparations:
>  - 6 nodes with OS CentOs7.7.1908, Ceph14.2.8:
>  - cephtest01,cephtest02,cephtest03: mon+mgr+mds+rgw
>  - cephtest04,cephtest05,cephtest06: osd disks (for test purposes only one disk by node)
>  - default pools for rgw
>  - replicated pool for metadata and erasured pool for data For CephFs
> 
> 0) Set flag noout
> # ceph osd set noout
> 1) I made a backup of admin keyring and config from /etc/ceph/ (it must be equal on all nodes)
> 2) I stopped all services on cephtest03 and remembered their names
> ceph-mon@cephtest03 
> ceph-mds@cephtest03 
> ceph-mgr@cephtest03 
> ceph-radosgw@rgw.cephtest03 
> 3) Then made a full backup of  /var/lib/ceph/
> 4) A new OS (Debian 10) is installed with the same host name and ip
> 5) New packages for Ceph15.2.0 installed
> 6) I copied admin keyring and config back to /etc/ceph/
> 7) Returned everything to /var/lib/ceph/
> 8) The monitor seems to need rights, so I did
> # chown -R ceph:ceph /var/lib/ceph/mon
> 9) Then I turned on the same services that were before in 2)
> # systemctl enable ceph-mon@cephtest03
> # systemctl enable ceph-mds@cephtest03
> # systemctl enable ceph-mgr@cephtest03
> # systemctl enable ceph-radosgw@rgw.cephtest03
> 10) Started only the monitor
> # systemctl start  ceph-mon@cephtest03
> 11) Wait until a quorum is formed
> 12) Started other services
> # systemctl start  ceph-mds@cephtest03
> # systemctl start   ceph-mgr@cephtest03
> # systemctl start  ceph-radosgw@rgw.cephtest03
> 13) Repeat steps 2-12 with cephtest02 and cephtest01
> 14) Now the next part - osd. Started with cephtest04
> 15) It seems that I only need to remember names such as
> ceph-volume@lvm-{osd_id}-{lv_name}
> ceph-osd@{osd_id}
> 16) I stopped all services and made sure that I have all pgs in the active state+clean (and some of them in the active+degraded)
> 17) I saved only keyring from /var/lib/ceph/bootstrap-osd/
> 18) steps 4-6
> 19) Returned bootstrap-osd keyring from 17) back to /var/lib/ceph/bootstrap-osd/ 
> 20) Then I turned on the same services that were before at 15) and started ceph-volume
> # systemctl enable ceph-volume@lvm-{osd_id}-{lv_name}
> # systemctl enable --runtime ceph-osd@{osd_id}
> # systemctl start  ceph-volume@lvm-{osd_id}-{lv_name}
> 23) Repeat steps 15-22 with cephtest05 and cephtest06
> 24) Unset flag noout
> # ceph osd unset noout
> 
> I got an upgraded cluster with working dashboard and HEALTH_OK :)
> And I want to ask only one question: am I missing something or not, because it seems pretty easy to switch from one hosting OS to another?
wow, this sounds promising. Thanks for the detailed workflow!
In our case it will be even easier since when we planned our first
installation we put /var/lib/ceph on a separate partition. So no copying
needed.

Dietmar

-- 
_________________________________________
D i e t m a r  R i e d e r, Mag.Dr.
Innsbruck Medical University
Biocenter - Institute of Bioinformatics
Innrain 80, 6020 Innsbruck
Email: dietmar.rieder@xxxxxxxxxxx
Web:   http://www.icbi.at
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux