/dev/mapper/mpathc is not mounted. 'file -c' output just shows the below. # mkqdisk -f qdisk0 mkqdisk v3.0.12.1 /dev/block/253:3: /dev/disk/by-id/dm-name-mpathc: /dev/disk/by-id/dm-uuid-mpath-36005076802810f60500000000000000f: /dev/dm-3: /dev/mapper/mpathc: Magic: eb7a62c2 Label: qdisk0 Created: Thu Feb 13 14:22:52 2014 Host: HITGSMQ02 Kernel Sector Size: 512 Recorded Sector Size: 512 # file -c /dev/mapper/mpathc cont offset type opcode mask value desc # -----Original Message----- From: discuss-bounces@xxxxxxxxxxxx [mailto:discuss-bounces@xxxxxxxxxxxx] On Behalf Of Christine Caulfield Sent: Thursday, November 27, 2014 5:54 PM To: discuss@xxxxxxxxxxxx Subject: Re: Desperate for Help - Cluster Node randomly reboots On 27/11/14 09:50, Tan Ban Wee wrote: > Thanks. > > How can I related the below label of the qdisk to the actual disk that the OS is seeing? > > <quorumd label="qdisk0"> > <heuristic interval="3" program="ping -c3 -w5 10.101.210.250" tko="5"/> > </quorumd> > > mkqdisk -f qdisk0 Chrissie > > -----Original Message----- > From: discuss-bounces@xxxxxxxxxxxx [mailto:discuss-bounces@xxxxxxxxxxxx] On Behalf Of Christine Caulfield > Sent: Thursday, November 27, 2014 5:14 PM > To: discuss@xxxxxxxxxxxx > Subject: Re: Desperate for Help - Cluster Node randomly reboots > > The first thing is to check your qdisk, as that's the daemon that's > causing the reboots. it's complaining about the CRC while reading the > header which leads me to think that the drive/partition is either > offline or corrupted. As the contents are consistent and "0x190a55ad" > then it seems most likely to me that the partition is corrupt - perhaps > formatted as a filesystem by mistake or something else has gone badly wrong. > > So. check the partition is OK and not mounted. 'file -c' will tell you > if it's been formatted as something else, for a qdisk it'll just come > back with 'data' as it's mostly zeroes - which is also something to check. > > Recreating the qdisk with mkqdisk -l <label> -f <device> might resurrect > it in this case, and is well worth trying. but make sure it's not a > mounted filesystem first as it'll just get corrupted again! > > Chrissie > > > On 27/11/14 08:52, Tan Ban Wee wrote: >> Hi, >> >> Thanks for replying. Is there any logs that I can provide so that we can have more leads to the cause? >> >> Below are from messages. Node 2 got evicted and reboots at 22:37. >> >> These random reboots happens almost everday. >> >> Nov 26 22:37:01 HITGSMQ01 rgmanager[31311]: [script] Executing /etc/init.d/SGQM04 status >> Nov 26 22:37:11 HITGSMQ01 rgmanager[31713]: [script] Executing /etc/init.d/SGQM03 status >> Nov 26 22:37:31 HITGSMQ01 rgmanager[32270]: [script] Executing /etc/init.d/SGQM04 status >> Nov 26 22:37:47 HITGSMQ01 qdiskd[2357]: Writing eviction notice for node 2 >> Nov 26 22:37:48 HITGSMQ01 qdiskd[2357]: Node 2 evicted >> Nov 26 22:37:51 HITGSMQ01 rgmanager[433]: [script] Executing /etc/init.d/SGQM03 status >> Nov 26 22:37:52 HITGSMQ01 corosync[2303]: [TOTEM ] A processor failed, forming new configuration. >> Nov 26 22:37:54 HITGSMQ01 corosync[2303]: [QUORUM] Members[1]: 1 >> Nov 26 22:37:54 HITGSMQ01 corosync[2303]: [TOTEM ] A processor joined or left the membership and a new membership was forme >> d. >> Nov 26 22:37:54 HITGSMQ01 kernel: dlm: closing connection to node 2 >> Nov 26 22:37:54 HITGSMQ01 rgmanager[3146]: State change: HITGSMQ02-hb DOWN >> Nov 26 22:37:54 HITGSMQ01 corosync[2303]: [CPG ] chosen downlist: sender r(0) ip(10.1.3.3) ; members(old:2 left:1) >> Nov 26 22:37:54 HITGSMQ01 corosync[2303]: [MAIN ] Completed service synchronization, ready to provide service. >> Nov 26 22:39:57 HITGSMQ01 kernel: INFO: task rgmanager:469 blocked for more than 120 seconds. >> Nov 26 22:39:57 HITGSMQ01 kernel: "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. >> Nov 26 22:39:57 HITGSMQ01 kernel: rgmanager D 0000000000000000 0 469 3144 0x00000080 >> Nov 26 22:39:57 HITGSMQ01 kernel: ffff880c24111c60 0000000000000086 ffff880c24111c28 ffff880c24111c24 >> Nov 26 22:39:57 HITGSMQ01 kernel: ffffffff81055f76 ffff880c7fc23080 ffff880028316700 000000000000047e >> Nov 26 22:39:57 HITGSMQ01 kernel: ffff880c692ad058 ffff880c24111fd8 000000000000fb88 ffff880c692ad058 >> Nov 26 22:39:57 HITGSMQ01 kernel: Call Trace: >> Nov 26 22:39:57 HITGSMQ01 kernel: [<ffffffff81055f76>] ? enqueue_task+0x66/0x80 >> Nov 26 22:39:57 HITGSMQ01 kernel: [<ffffffff8150fc45>] rwsem_down_failed_common+0x95/0x1d0 >> Nov 26 22:39:57 HITGSMQ01 kernel: [<ffffffff8150fdd6>] rwsem_down_read_failed+0x26/0x30 >> Nov 26 22:39:57 HITGSMQ01 kernel: [<ffffffff812833b4>] call_rwsem_down_read_failed+0x14/0x30 >> Nov 26 22:39:57 HITGSMQ01 kernel: [<ffffffff8150f2d4>] ? down_read+0x24/0x30 >> Nov 26 22:39:57 HITGSMQ01 kernel: [<ffffffffa038e257>] dlm_user_request+0x47/0x1b0 [dlm] >> Nov 26 22:39:57 HITGSMQ01 kernel: [<ffffffff8106659b>] ? dequeue_task_fair+0x12b/0x130 >> Nov 26 22:39:57 HITGSMQ01 kernel: [<ffffffff81167c53>] ? kmem_cache_alloc_trace+0x1a3/0x1b0 >> Nov 26 22:39:57 HITGSMQ01 kernel: [<ffffffffa039b2a7>] device_write+0x5c7/0x720 [dlm] >> Nov 26 22:39:57 HITGSMQ01 kernel: [<ffffffff8109c424>] ? switch_task_namespaces+0x24/0x60 >> Nov 26 22:39:57 HITGSMQ01 kernel: [<ffffffff8121baf6>] ? security_file_permission+0x16/0x20 >> Nov 26 22:39:57 HITGSMQ01 kernel: [<ffffffff81180f98>] vfs_write+0xb8/0x1a0 >> Nov 26 22:39:57 HITGSMQ01 kernel: [<ffffffff81181891>] sys_write+0x51/0x90 >> Nov 26 22:39:57 HITGSMQ01 kernel: [<ffffffffa03b09df>] twnotify_sys_write+0x1f/0x80 [twnotify] >> Nov 26 22:39:57 HITGSMQ01 kernel: [<ffffffff8100b072>] system_call_fastpath+0x16/0x1b >> Nov 26 22:40:12 HITGSMQ01 corosync[2303]: [TOTEM ] A processor joined or left the membership and a new membership was for >> >> -----Original Message----- >> From: discuss-bounces@xxxxxxxxxxxx [mailto:discuss-bounces@xxxxxxxxxxxx] On Behalf Of Christine Caulfield >> Sent: Thursday, November 27, 2014 4:40 PM >> To: discuss@xxxxxxxxxxxx >> Subject: Re: Desperate for Help - Cluster Node randomly reboots >> >> On 26/11/14 09:24, Tan Ban Wee wrote: >>> Hi, >>> >>> This is a 2 nodes cluster and they are randomly rebooting itself. I hope >>> someone can help me to narrow down to the cause. >>> >>> Nov 25 23:40:17 qdiskd Node 1 missed an update (3/4) >>> >>> Nov 25 23:40:18 qdiskd Node 1 missed an update (4/4) >>> >>> Nov 25 23:40:19 qdiskd Node 1 missed an update (5/4) >>> >>> Nov 25 23:40:19 qdiskd Node 1 DOWN >>> >>> Nov 25 23:40:19 qdiskd Writing eviction notice for node 1 >>> >>> Nov 25 23:40:19 qdiskd Telling CMAN to kill the node >>> >>> Nov 25 23:40:20 qdiskd Node 1 evicted >>> >>> Nov 25 23:44:19 qdiskd Node 1 is UP >>> >>> Nov 25 23:44:20 qdiskd Node 1 shutdown >>> >>> Nov 25 23:44:26 qdiskd Node 1 is UP >>> >>> Nov 25 23:44:37 qdiskd Node 1 missed an update (2/4) >>> >>> Nov 25 23:44:38 qdiskd Node 1 missed an update (3/4) >>> >>> Nov 25 23:44:39 qdiskd Node 1 missed an update (4/4) >>> >>> Nov 25 23:44:40 qdiskd Node 1 missed an update (5/4) >>> >>> Nov 25 23:44:40 qdiskd Node 1 DOWN >>> >>> Nov 25 23:44:40 qdiskd Writing eviction notice for node 1 >>> >>> Nov 25 23:44:40 qdiskd Telling CMAN to kill the node >>> >>> Nov 25 23:44:41 qdiskd Node 1 evicted >>> >>> Nov 25 23:50:48 qdiskd Loading dynamic configuration >>> >>> Nov 25 23:50:49 qdiskd Setting autocalculated votes to 1 >>> >>> Nov 25 23:50:49 qdiskd Loading static configuration >>> >>> Nov 25 23:50:49 qdiskd Auto-configured TKO as 4 based on token=10000 >>> interval=1 >>> >>> Nov 25 23:50:49 qdiskd Timings: 4 tko, 1 interval >>> >>> Nov 25 23:50:49 qdiskd Timings: 2 tko_up, 3 master_wait, 2 upgrade_wait >>> >>> Nov 25 23:50:49 qdiskd Heuristic: 'ping -c3 -w5 10.101.210.250' score=1 >>> interval=3 tko=5 >>> >>> Nov 25 23:50:49 qdiskd 1 heuristics loaded >>> >>> Nov 25 23:50:49 qdiskd Quorum Daemon: 1 heuristics, 1 interval, 4 tko, 1 >>> votes >>> >>> Nov 25 23:50:49 qdiskd Run Flags: 00000231 >>> >>> Nov 25 23:50:49 qdiskd Header CRC32 mismatch; Exp: 0x00000000 Got: >>> 0x190a55ad >>> >>> Nov 25 23:50:49 qdiskd diskRawReadShadow: bad CRC32, offset = 0 len = 512 >>> >>> Nov 25 23:50:49 qdiskd Header CRC32 mismatch; Exp: 0x00000000 Got: >>> 0x190a55ad >>> >>> Nov 25 23:50:49 qdiskd diskRawReadShadow: bad CRC32, offset = 0 len = 512 >>> >>> Nov 25 23:50:49 qdiskd Header CRC32 mismatch; Exp: 0x00000000 Got: >>> 0x190a55ad >>> >>> Nov 25 23:50:49 qdiskd diskRawReadShadow: bad CRC32, offset = 0 len = 512 >>> >>> Nov 25 23:50:49 qdiskd Header CRC32 mismatch; Exp: 0x00000000 Got: >>> 0x190a55ad >>> >>> Nov 25 23:50:49 qdiskd diskRawReadShadow: bad CRC32, offset = 0 len = 512 >>> >>> Nov 25 23:50:49 qdiskd Header CRC32 mismatch; Exp: 0x00000000 Got: >>> 0x190a55ad >>> >>> Nov 25 23:50:49 qdiskd diskRawReadShadow: bad CRC32, offset = 0 len = 512 >>> >>> Nov 25 23:50:49 qdiskd Header CRC32 mismatch; Exp: 0x00000000 Got: >>> 0x190a55ad >>> >> >> From that limited information I would guess that your quorum disk >> partition is either offline or corrupted. First check that the drive is >> online and if it seems OK physically then check that it's not been >> formatted as a filesystem or something else by mistake and rebuild the >> header using mkqdisk. >> >> Chrissie >> >> >> _______________________________________________ >> discuss mailing list >> discuss@xxxxxxxxxxxx >> http://lists.corosync.org/mailman/listinfo/discuss >> This email message may contain confidential and privileged information for its intended recipient(s) only. If you are not the intended recipient or the addressee, you may have received it by unauthorised means. You are to notify the sender immediately and thereafter delete the email. You are to take note that any disclosure, distribution, use or storage of this communication is strictly prohibited. Any opinions, conclusions and other information in this message that are unrelated to official business of the RHB Banking Group are those of the individual sender and shall be understood as neither explicitly given nor endorsed by the RHB Banking Group. RHB Banking Group shall not be liable for any loss or damage caused by viruses transmitted by this e-mail or its attachments. Further, the RHB Banking Group is also not responsible for any unauthorised changes made to the information or the effect thereto >> >> _______________________________________________ >> discuss mailing list >> discuss@xxxxxxxxxxxx >> http://lists.corosync.org/mailman/listinfo/discuss >> > > _______________________________________________ > discuss mailing list > discuss@xxxxxxxxxxxx > http://lists.corosync.org/mailman/listinfo/discuss > This email message may contain confidential and privileged information for its intended recipient(s) only. If you are not the intended recipient or the addressee, you may have received it by unauthorised means. You are to notify the sender immediately and thereafter delete the email. You are to take note that any disclosure, distribution, use or storage of this communication is strictly prohibited. Any opinions, conclusions and other information in this message that are unrelated to official business of the RHB Banking Group are those of the individual sender and shall be understood as neither explicitly given nor endorsed by the RHB Banking Group. RHB Banking Group shall not be liable for any loss or damage caused by viruses transmitted by this e-mail or its attachments. Further, the RHB Banking Group is also not responsible for any unauthorised changes made to the information or the effect thereto > _______________________________________________ discuss mailing list discuss@xxxxxxxxxxxx http://lists.corosync.org/mailman/listinfo/discuss This email message may contain confidential and privileged information for its intended recipient(s) only. If you are not the intended recipient or the addressee, you may have received it by unauthorised means. You are to notify the sender immediately and thereafter delete the email. You are to take note that any disclosure, distribution, use or storage of this communication is strictly prohibited. Any opinions, conclusions and other information in this message that are unrelated to official business of the RHB Banking Group are those of the individual sender and shall be understood as neither explicitly given nor endorsed by the RHB Banking Group. RHB Banking Group shall not be liable for any loss or damage caused by viruses transmitted by this e-mail or its attachments. Further, the RHB Banking Group is also not responsible for any unauthorised changes made to the information or the effect thereto _______________________________________________ discuss mailing list discuss@xxxxxxxxxxxx http://lists.corosync.org/mailman/listinfo/discuss