Re: problem after growing

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Eric Sandeen <sandeen@xxxxxxxxxxx> a écrit :

>On 2/13/13 1:50 PM, Eric Sandeen wrote:
>> On 2/13/13 12:09 PM, Rémi Cailletaud wrote:
>>> Le 13/02/2013 18:52, Eric Sandeen a écrit :
>> 
>> <snip>
>> 
>>>> Did the filesystem grow actually work?
>>>>
>>>> # xfs_db -c "sb 0" -c "p" /dev/vg0/tomo-201111
>>>> magicnum = 0x58465342
>>>> blocksize = 4096
>>>> dblocks = 10468982745
>>>>
>>>> That looks like it's (still?) a 38TiB/42TB filesystem, with:
>>>>
>>>> sectsize = 512
>>>>
>>>> 512 sectors.
>>>>
>>>> How big was it before you tried to grow it, and how much did you
>try to grow it by?  Maybe the size never changed.
>>>
>>> Was 39, growing to 44. Testdisk says 48 TB / 44 TiB... There is some
>chance that it was never really growed.
>>>> At mount time it tries to set the sector size of the device; its' a
>hard-4k device, so setting it to 512 fails.
>>>>
>>>> This may be as much of an LVM issue as anything; how do you get the
>LVM device back to something with 512-byte logical sectors?  I have no
>idea...
>>>>
>>>> *if* the fs didn't actually grow, and if the new 4k-sector space is
>not used by the filesystem, and if you can somehow remove that new
>space from the device and set the LV back to 512 sectors, you might be
>in good shape.
>>> I dont either know how to see nor set LV sector size.  It's 100%
>sure that anything was copied on 4k sector size, and pretty sure that
>the fs did not really grow.
>> 
>> I think the same blockdev command will tell you.
>> 
>>  
>>>> Proceed with extreme caution here, I wouldn't start just trying
>random things unless you have some other way to get your data back
>(backups?).  I'd check with LVM folks as well, and maybe see if
>dchinner or the sgi folks have other suggestions.
>>> Sigh... No backup (44To is too large for us...) ! I'm running a
>testdisk recover, but I'm not very confident about success...
>>> Thanks to deeper investigate this...
>>>> First let's find out if the filesystem actually thinks it's living
>on the new space.
>>> What is the way to make it talk about that ?
>> 
>> well, you have 10468982745 4k blocks in your filesystem, so
>42880953323520 bytes of xfs filesystem.
>> 
>> Look at your lvm layout, does that extend into the new disk space or
>is it confined to the original disk space?
>
>lvm folks I talk to say that if you remove the 4k device from the lvm
>volume it should switch back to 512 sectors.
>
>so if you can can convince yourself that 42880953323520 bytes does not
>cross into the newly added disk space, just remove it again, and
>everything should be happy.
>

Ok, I'll check that tomorrow and try to remove added space if my fs does not live on...

>Unless your rash decision to start running "testdisk" made things worse
>;)

It's only analyse, it does not modify anything... it should not have any effect...

Thx again,

rémi

>-Eric
>
>> -Eric
>> 
>>> Thanks again for your help !
>>>
>>> rémi
>>>
>>>> -Eric
>>>>
>>>>> rémi
>
>_______________________________________________
>xfs mailing list
>xfs@xxxxxxxxxxx
>http://oss.sgi.com/mailman/listinfo/xfs


-- 
Envoyé de mon téléphone Android avec K-9 Mail. Excusez la brièveté.

_______________________________________________
xfs mailing list
xfs@xxxxxxxxxxx
http://oss.sgi.com/mailman/listinfo/xfs



[Index of Archives]     [Linux XFS Devel]     [Linux Filesystem Development]     [Filesystem Testing]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux