Re: CephFS ghost usage/inodes

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



i run this commands, but still the same problems

$ cephfs-data-scan scan_extents cephfs_data

$ cephfs-data-scan scan_inodes cephfs_data

$ cephfs-data-scan scan_links
2020-01-14 20:36:45.110 7ff24200ef80 -1 mds.0.snap  updating last_snap 1
-> 27

$ cephfs-data-scan cleanup cephfs_data

do you have other ideas ?

Am 14.01.20 um 20:32 schrieb Patrick Donnelly:
> On Tue, Jan 14, 2020 at 11:24 AM Oskar Malnowicz
> <oskar.malnowicz@xxxxxxxxxxxxxx> wrote:
>> $ ceph daemon mds.who flush journal
>> {
>>     "message": "",
>>     "return_code": 0
>> }
>>
>>
>> $ cephfs-table-tool 0 reset session
>> {
>>     "0": {
>>         "data": {},
>>         "result": 0
>>     }
>> }
>>
>> $ cephfs-table-tool 0 reset snap
>> {
>>     "result": 0
>> }
>>
>> $ cephfs-table-tool 0 reset inode
>> {
>>     "0": {
>>         "data": {},
>>         "result": 0
>>     }
>> }
>>
>> $ cephfs-journal-tool --rank=cephfs_test1:0 journal reset
>> old journal was 98282151365~92872
>> new journal start will be 98285125632 (2881395 bytes past old end)
>> writing journal head
>> writing EResetJournal entry
>> done
>>
>> $ cephfs-data-scan init
>> Inode 0x0x1 already exists, skipping create.  Use --force-init to
>> overwrite the existing object.
>> Inode 0x0x100 already exists, skipping create.  Use --force-init to
>> overwrite the existing object.
>>
>> Should i run with --force-init flag ?
> No, that shouldn't be necessary.
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux