Re: Recovery question

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



-----BEGIN PGP SIGNED MESSAGE-----
Hash: SHA256

The default is /var/lib/ceph/mon/- (/var/lib/ceph/mon/ceph-mon1 for
me). You will also need the information from /etc/ceph/ to reconstruct
the data. I *think* you should be able to just copy this to a new box
with the same name and IP address and start it up.

I haven't actually done this, so there still may be some bumps.
- ----------------
Robert LeBlanc
PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62 B9F1


On Wed, Jul 29, 2015 at 3:44 PM, Peter Hinman  wrote:
> Thanks Robert -
>
> Where would that monitor data (database) be found?
>
> --
> Peter Hinman
>
>
> On 7/29/2015 3:39 PM, Robert LeBlanc wrote:
>>
>> -----BEGIN PGP SIGNED MESSAGE-----
>> Hash: SHA256
>>
>> If you built new monitors, this will not work. You would have to
>> recover the monitor data (database) from at least one monitor and
>> rebuild the monitor. The new monitors would not have any information
>> about pools, OSDs, PGs, etc to allow an OSD to be rejoined.
>> - ----------------
>> Robert LeBlanc
>> PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62 B9F1
>>
>>
>> On Wed, Jul 29, 2015 at 2:46 PM, Peter Hinman  wrote:
>>>
>>> Hi Greg -
>>>
>>> So at the moment, I seem to be trying to resolve a permission error.
>>>
>>>   === osd.3 ===
>>>   Mounting xfs on stor-2:/var/lib/ceph/osd/ceph-3
>>>   2015-07-29 13:35:08.809536 7f0a0262e700  0 librados: osd.3
>>> authentication
>>> error (1) Operation not permitted
>>>   Error connecting to cluster: PermissionError
>>>   failed: 'timeout 30 /usr/bin/ceph -c /etc/ceph/ceph.conf --name=osd.3
>>> --keyring=/var/lib/ceph/osd/ceph-3/keyring osd crush create-or-move -- 3
>>> 3.64 host=stor-2 root=default'
>>>   ceph-disk: Error: ceph osd start failed: Command '['/usr/sbin/service',
>>> 'ceph', '--cluster', 'ceph', 'start', 'osd.3']' returned non-zero exit
>>> status 1
>>>   ceph-disk: Error: One or more partitions failed to activate
>>>
>>>
>>> Is there a way to identify the cause of this PermissionError?  I've
>>> copied
>>> the client.bootstrap-osd key from the output of ceph auth list, and
>>> pasted
>>> it into /var/lib/ceph/bootstrap-osd/ceph.keyring, but that has not
>>> resolve
>>> the error.
>>>
>>> But it sounds like you are saying that even once I get this resolved, I
>>> have
>>> no hope of recovering the data?
>>>
>>> --
>>> Peter Hinman
>>>
>>> On 7/29/2015 1:57 PM, Gregory Farnum wrote:
>>>
>>> This sounds like you're trying to reconstruct a cluster after destroying
>>> the
>>> monitors. That is...not going to work well. The monitors define the
>>> cluster
>>> and you can't move OSDs into different clusters. We have ideas for how to
>>> reconstruct monitors and it can be done manually with a lot of hassle,
>>> but
>>> the process isn't written down and there aren't really fools I help with
>>> it.
>>> :/
>>> -Greg
>>>
>>> On Wed, Jul 29, 2015 at 5:48 PM Peter Hinman  wrote:
>>>>
>>>> I've got a situation that seems on the surface like it should be
>>>> recoverable, but I'm struggling to understand how to do it.
>>>>
>>>> I had a cluster of 3 monitors, 3 osd disks, and 3 journal ssds. After
>>>> multiple hardware failures, I pulled the 3 osd disks and 3 journal ssds
>>>> and am attempting to bring them back up again on new hardware in a new
>>>> cluster.  I see plenty of documentation on how to zap and initialize and
>>>> add "new" osds, but I don't see anything on rebuilding with existing osd
>>>> disks.
>>>>
>>>> Could somebody provide guidance on how to do this?  I'm running 94.2 on
>>>> all machines.
>>>>
>>>> Thanks,
>>>>
>>>> --
>>>> Peter Hinman
>>>>
>>>>
>>>> _______________________________________________
>>>> ceph-users mailing list
>>>> ceph-users@xxxxxxxxxxxxxx
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>>
>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@xxxxxxxxxxxxxx
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>> -----BEGIN PGP SIGNATURE-----
>> Version: Mailvelope v0.13.1
>> Comment: https://www.mailvelope.com
>>
>> wsFcBAEBCAAQBQJVuUfwCRDmVDuy+mK58QAASJIP/0kEBx+h7LZfpQkgUvPG
>> hKzIlSzbWIkig9O5cYzXKh03jPFz1hj38YVQ+cdYuRA1VhrNdTkwyNnzVDFk
>> 7R98PUF4eljNNnSdQ0nIIVCS8rtGWfSUU4ECo1/4Gm8ebIMmY/g6umE87oqy
>> fBmXW9luFZ3HQyoaqfALKWsesNJ9EJT/EgMH3+XisJZYPtpEbVDr0DiV2sbt
>> st1xtsQwkKOGAOr+7sGe7g9dED7zCERLWsNOpeHkeJaArbKDzGY1abpoiyUt
>> BQ5lCHGKZCBqXINaVTmwPGMTdKpED5eBxIXQ+QeEXwBONQuei4zkDz8TWRKO
>> zaNcogcaQilSg3KyjyHzovPzVoS0OGLmEK1FVtveUfMPfMQ9XXyGnhWiZ6u7
>> +grlQoe4E5AZTqEMtCzKyrqldWdzL8A+S9ZidtvSi1dCZpJutEkFbI/m8A5j
>> dA6Q7zijNJDPVMMsXXA08z6Pu7611mShXjW0fLu871++JsE/eS8GCfc9Cgyu
>> aUgcSaWCuRVa2laXak3BI+44AexsU3ZKyveDeuFdm7y3F+DS5FKZK2V8OfJn
>> /mbolRFyGCaBEj83FQJGCBrsSOzYDhas8aEDa4W9kKLbKeBaeRUE0mXQYfvu
>> 12lZxpzn0UasrH/mcgu8ij9ElLN5Fq0wSp1SNKbg/RczcYVt/DjjGbCRDTgO
>> b23I
>> =1NQh
>> -----END PGP SIGNATURE-----
>
>
>

-----BEGIN PGP SIGNATURE-----
Version: Mailvelope v0.13.1
Comment: https://www.mailvelope.com

wsFcBAEBCAAQBQJVuUnoCRDmVDuy+mK58QAAM3MP/jcJUwKHRUjI47y+S1St
5Tmkpmob0CdaplzdTvTYF9nc0U6GxhEwTxs/e6CuJ8raKKCNdV2Bu303kQQz
t9xAJXbzUe4w5/f3WwlM91R3JavcqRVKyiVptzRf3EuHMCWUmLhWKlOK6WNF
fcO3yRRyn5gkGHR2w7MAAVvABSyZsrLSvCU8Ckila3/cdZNRkOUCfiwSfYlL
TXd1fsOVP629ltvl5xFQ2u4Tmx+hHq+oxx9PesWz8eKbXwM0Cbq84WfwojlQ
GHL9DZlCOSGwsinEHQycODGmkL7QSfEBYEZYbMgtDmUbUOXLfUifvKvuk7yD
I1QxzjhWLxa0k4k/8YzP8055euFTIzZ3E4g3jG50KzvxIaUhHuiGT0CXEtEi
vEXg1iGtY3lhNjoMNMjQcSbi/PlbBL+08Y6y/VboXu4Tk6jm62g1gJvBHhUE
OuE3Yiw1wQZI8LjGpCHQOt/ck3aFotsQI73hZF2DwANFeEKKcZvi+Pj0QsdC
XVTCjI4qM4k36HjJG3tT/94Y6PbRVyeK6bLCa9klmx/u4ToWUjgD5e7Nc6Pv
OcIkUfG+U6v5TjMAsGzEu1ZjQWNdpEGPoRr+cqwBYNlZu8QUacIAZ86jIyPe
77lrKrRbiYsdaYzZ48jrBXM+RiJY/sHgWnAr/Y12ysGbel3LVqxp8gGfI6zg
beQf
=gokk
-----END PGP SIGNATURE-----
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux