Hi all,
Since I sent the below, the aforementioned cluster crashed. Now I
can't mount the scratch112 filesystem. Attempts to do so crash the
node trying to mount it. If I run gfs_fsck against it, I see the
following:
# gfs_fsck -nv /dev/s12/scratch112
Initializing fsck
Initializing lists...
Initializing special inodes...
Validating Resource Group index.
Level 1 check.
5834 resource groups found.
(passed)
Setting block ranges...
Can't seek to last block in file system: 4969529913
Unable to determine the boundaries of the file system.
Freeing buffers.
Not being able to determine the boundaries of the file system seems
like a very bad thing. However, LVM didn't complain in the slightest
when I expanded the logical volume. How can I recover from this?
Thanks,
James
On Aug 29, 2008, at 9:19 PM, James Chamberlain wrote:
Hi all,
I'm trying to grow a GFS filesystem. I've grown this filesystem
before and everything went fine. However, when I issued gfs_grow
this time, I saw the following messages in my logs:
Aug 29 21:04:13 s12n02 kernel: lock_dlm: lm_dlm_cancel 2,17 flags 80
Aug 29 21:04:13 s12n02 kernel: lock_dlm: lm_dlm_cancel skip 2,17
flags 100
Aug 29 21:04:14 s12n02 kernel: lock_dlm: lm_dlm_cancel 2,17 flags 80
Aug 29 21:04:14 s12n02 kernel: dlm: scratch112: (14239) dlm_unlock:
10241 busy 2
Aug 29 21:04:14 s12n02 kernel: lock_dlm: lm_dlm_cancel rv -16 2,17
flags 40080
The last three lines of these log entries repeat themselves once a
second until I hit ^C. The filesystem appears to still be up and
accessible. Any thoughts on what's going on here and what I can do
about it?
Thanks,
James
--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster
--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster