Srinivas, Could you paste the output of "gluster volume info gvol1"? This should give us an idea as to what was the state of the volume before the power loss. thanks, krish ----- Original Message ----- > Hello Gluster users: > sorry for long post, I have run out of ideas here, kindly let me know if i am > looking at right places for logs and any suggested actions.....thanks > a sudden power loss casued hard reboot - now the volume does not start > Glusterfs- 3.3.1 on Centos 6.1 transport: TCP > sharing volume over NFS for VM storage - VHD Files > Type: distributed - only 1 node (brick) > XFS (LVM) > mount /dev/datastore1/mylv1 /export/brick1 - mounts VHD files.......is there > a way to recover these files? > cat export-brick1.log > [2013-06-02 09:29:00.832914] I [glusterfsd.c:1666:main] > 0-/usr/sbin/glusterfsd: Started running /usr/sbin/glusterfsd version 3.3.1 > [2013-06-02 09:29:00.845515] I [graph.c:241:gf_add_cmdline_options] > 0-gvol1-server: adding option 'listen-port' for volume 'gvol1-server' with > value '24009' > [2013-06-02 09:29:00.845558] I [graph.c:241:gf_add_cmdline_options] > 0-gvol1-posix: adding option 'glusterd-uuid' for volume 'gvol1-posix' with > value '16ee7a4e-ee9b-4543-bd61-9b444100693d' > [2013-06-02 09:29:00.846654] W [options.c:782:xl_opt_validate] > 0-gvol1-server: option 'listen-port' is deprecated, preferred is > 'transport.socket.listen-port', continuing with correction > Given volfile: > +------------------------------------------------------------------------------+ > 1: volume gvol1-posix > 2: type storage/posix > 3: option directory /export/brick1 > 4: option volume-id aa25aa58-d191-432a-a84b-325051347af6 > 5: end-volume > 6: > 7: volume gvol1-access-control > 8: type features/access-control > 9: subvolumes gvol1-posix > 10: end-volume > 11: > 12: volume gvol1-locks > 13: type features/locks > 14: subvolumes gvol1-access-control > ---------- > ----------------- > > 46: option transport-type tcp > 47: option auth.login./export/brick1.allow > 6c4653bb-b708-46e8-b3f9-177b4cdbbf28 > 48: option auth.login.6c4653bb-b708-46e8-b3f9-177b4cdbbf28.password > 091ae3b1-40c2-4d48-8870-6ad7884457ac > 49: option auth.addr./export/brick1.allow * > 50: subvolumes /export/brick1 > 51: end-volume > +------------------------------------------------------------------------------+ > [2013-06-02 09:29:03.963001] W [socket.c:410:__socket_keepalive] 0-socket: > failed to set keep idle on socket 8 > [2013-06-02 09:29:03.963046] W [socket.c:1876:socket_server_event_handler] > 0-socket.glusterfsd: Failed to set keep-alive: Operation not supported > [2013-06-02 09:29:04.850120] I [server-handshake.c:571:server_setvolume] > 0-gvol1-server: accepted client from > iiclab-oel1-9347-2013/06/02-09:29:00:835397-gvol1-client-0-0 (version: > 3.3.1) > [2013-06-02 09:32:16.973786] W [glusterfsd.c:831:cleanup_and_exit] > (-->/usr/lib64/libgfrpc.so.0(rpcsvc_notify+0x93) [0x30cac0a5b3] > (-->/usr/lib64/libgfrpc.so.0(rpcsvc_handle_rpc_call+0x293) [0x30cac0a443] > (-->/usr/sbin/glusterfsd(glusterfs_handle_terminate+0x15) [0x40a955]))) 0-: > received signum (15), shutting down > [2013-06-02 09:32:16.973895] W [glusterfsd.c:831:cleanup_and_exit] > (-->/lib64/libc.so.6(clone+0x6d) [0x3ef56e68ed] (-->/lib64/libpthread.so.0() > [0x3ef5a077e1] (-->/usr/sbin/glusterfsd(glusterfs_sigwaiter+0xdd) > [0x405d4d]))) 0-: received signum (15), shutting down > NFS LOG > [2013-06-02 09:29:00.918906] I [rpc-clnt.c:1657:rpc_clnt_reconfig] > 0-gvol1-client-0: changing port to 24009 (from 0) > [2013-06-02 09:29:03.963023] W [socket.c:410:__socket_keepalive] 0-socket: > failed to set keep idle on socket 8 > [2013-06-02 09:29:03.963062] W [socket.c:1876:socket_server_event_handler] > 0-socket.glusterfsd: Failed to set keep-alive: Operation not supported > [2013-06-02 09:29:04.849941] I > [client-handshake.c:1636:select_server_supported_programs] 0-gvol1-client-0: > Using Program GlusterFS 3.3.1, Num (1298437), Version (330) > [2013-06-02 09:29:04.853016] I [client-handshake.c:1433:client_setvolume_cbk] > 0-gvol1-client-0: Connected to 10.0.0.30:24009, attached to remote volume > '/export/brick1'. > [2013-06-02 09:29:04.853048] I [client-handshake.c:1445:client_setvolume_cbk] > 0-gvol1-client-0: Server and Client lk-version numbers are not same, > reopening the fds > [2013-06-02 09:29:04.853262] I > [client-handshake.c:453:client_set_lk_version_cbk] 0-gvol1-client-0: Server > lk version = 1 > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://supercolony.gluster.org/mailman/listinfo/gluster-users