Re: Adding a host node back to ceph cluster

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Thank you Marc,

I wasn't aware that was an option so it will be very useful in future.
I see for Ubuntu you can make use of debsums to verify packages.

Sadly I'm still looking for a solution to my host issue though.


Kind regards
Geoffrey Rhodes

On 15 January 2018 at 23:31, Marc Roos <M.Roos@xxxxxxxxxxxxxxxxx> wrote:

Maybe for the future:

rpm {-V|--verify} [select-options] [verify-options]

       Verifying a package compares information about the installed
files in the package with information about  the  files  taken
       from  the  package  metadata  stored in the rpm database.  Among
other things, verifying compares the size, digest, permis‐
       sions, type, owner and group of each file.  Any discrepancies are
displayed.  Files that were not installed from the  pack‐
       age, for example, documentation files excluded on installation
using the "--excludedocs" option, will be silently ignored.





-----Original Message-----
From: Geoffrey Rhodes [mailto:geoffrey@xxxxxxxxxxxxx]
Sent: maandag 15 januari 2018 16:39
To: ceph-users@xxxxxxxxxxxxxx
Subject: Adding a host node back to ceph cluster

Good day,

I'm having an issue re-deploying a host back into my production ceph
cluster.
Due to some bad memory (picked up by a scrub) which has been replaced I
felt the need to re-install the host to be sure no host files were
damaged.

Prior to decommissioning the host I set the crush weight's on each osd
to 0.
Once to osd's had flushed all data I stopped the daemon's.
I then purged the osd's from the crushmap with "ceph osd purge".
Followed by "ceph osd crush rm {host}" to remove the host bucket from
the crush map.

I also ran "ceph-deploy purge {host}" & "ceph-deploy purgedata {host}"
from the management node.
I then reinstalled the host and made the necessary config changes
followed by the appropriate ceph-deploy commands (ceph-deploy
install..., ceph-deploy admin..., ceph-deploy osd create...) to bring
the host & it's osd's back into the cluster, - same as I would when
adding a new host node to the cluster.

Running ceph osd df tree shows the osd's however the host node is not
displayed.
Inspecting the crush map I see no host bucket has been created or any
host's osd's listed.
The osd's also did not start which explains the weight being 0 but I
presume the osd's not starting isn't the only issue since the crush map
lacks the newly installed host detail.

Could anybody maybe tell me where I've gone wrong?
I'm also assuming there shouldn't be an issue using the same host name
again or do I manually add the host bucket and osd detail back into the
crush map or should ceph-deploy not take care of that?

Thanks

OS: Ubuntu 16.04.3 LTS
Ceph version: 12.2.1 / 12.2.2 - Luminous


Kind regards
Geoffrey Rhodes




_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux