Re: Rebuilding/recreating CephFS journal?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



What's the current full output of "ceph -s"?

If you already had your MDS in damaged state, you might just need to
mark it as repaired. That's a monitor command.

On Fri, May 27, 2016 at 2:09 PM, Stillwell, Bryan J
<Bryan.Stillwell@xxxxxxxxxxx> wrote:
> On 5/27/16, 3:01 PM, "Gregory Farnum" <gfarnum@xxxxxxxxxx> wrote:
>
>>>
>>> So would the next steps be to run the following commands?:
>>>
>>> cephfs-table-tool 0 reset session
>>> cephfs-table-tool 0 reset snap
>>> cephfs-table-tool 0 reset inode
>>> cephfs-journal-tool --rank=0 journal reset
>>> cephfs-data-scan init
>>>
>>> cephfs-data-scan scan_extents data
>>> cephfs-data-scan scan_inodes data
>>
>>No, definitely not. I think you just need to reset the journal again,
>>since you wiped out a bunch of its data with that fs reset command.
>>Since your backing data should already be consistent you don't need to
>>do any data scans. Your snap and inode tables might be corrupt,
>>but...hopefully not. If they are busted...actually, I don't remember;
>>maybe you will need to run the data scan tooling to repair those. I'd
>>try to avoid it if possible just because of the time involved. (It'll
>>become obvious pretty quickly if the inode tables are no good.)
>
> So when I attempt to reset the journal again I get this:
>
> # cephfs-journal-tool journal reset
> journal does not exist on-disk. Did you set a bad rank?2016-05-27
> 15:03:30.016326 7f63f987e700  0 client.20626476.journaler(ro) error
> getting journal off disk
>
> Error loading journal: (2) No such file or directory, pass --force to
> forcibly reset this journal
> Error ((2) No such file or directory)
>
>
>
> And then I tried to force it which seemed to succeed:
>
> # cephfs-journal-tool journal reset --force
> writing EResetJournal entry
>
>
>
> However, when I restart the mds it gets stuck in standby mode:
>
> 2016-05-27 15:05:57.080672 7fe0cccd8700 -1 mds.b4 *** got signal
> Terminated ***
> 2016-05-27 15:05:57.080703 7fe0cccd8700  1 mds.b4 suicide.  wanted state
> up:standby
> 2016-05-27 15:06:04.527203 7f500f28a180  0 set uid:gid to 64045:64045
> (ceph:ceph)
> 2016-05-27 15:06:04.527259 7f500f28a180  0 ceph version 10.2.0
> (3a9fba20ec743699b69bd0181dd6c54dc01c64b9), process ceph-mds, pid 19163
> 2016-05-27 15:06:04.527569 7f500f28a180  0 pidfile_write: ignore empty
> --pid-file
> 2016-05-27 15:06:04.637842 7f5008a04700  1 mds.b4 handle_mds_map standby
>
>
>
> The relevant output from 'ceph -s' looks like this:
>
>       fsmap e287: 0/1/1 up, 1 up:standby, 1 damaged
>
>
> What am I missing?
>
> Thanks,
> Bryan
>
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux