Re: drbd

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]



On 12/10/14 02:30 AM, John R Pierce wrote:
so I've had a drbd replica running for a while of a 16TB raid thats used
as a backuppc repository.

when I have rebooted the backuppc server, the replica doesn't seem to
auto-restart til I do it manually, and the backupc /data file system on
this 16TB LUN doesn't seem to automount, either.

I've rebooted this thing a few times in the 18 months or so its been
running...   not always cleanly...

anyways, I'm started a drbd verify (from the slave) about 10 hours ago,
it has 15 hours more to run, and so far it's logged...

Oct 11 13:58:26 sg2 kernel: block drbd0: Starting Online Verify from
sector 3534084704
Oct 11 14:00:23 sg2 kernel: block drbd0: [drbd0_worker/2197]
sock_sendmsg time expired, ko = 4294967295
Oct 11 14:00:29 sg2 kernel: block drbd0: [drbd0_worker/2197]
sock_sendmsg time expired, ko = 4294967294
Oct 11 14:00:35 sg2 kernel: block drbd0: [drbd0_worker/2197]
sock_sendmsg time expired, ko = 4294967293
Oct 11 14:00:41 sg2 kernel: block drbd0: [drbd0_worker/2197]
sock_sendmsg time expired, ko = 4294967292
Oct 11 14:01:16 sg2 kernel: block drbd0: [drbd0_worker/2197]
sock_sendmsg time expired, ko = 4294967295
Oct 11 14:02:05 sg2 kernel: block drbd0: [drbd0_worker/2197]
sock_sendmsg time expired, ko = 4294967295
Oct 11 14:02:11 sg2 kernel: block drbd0: [drbd0_worker/2197]
sock_sendmsg time expired, ko = 4294967294
Oct 11 14:02:17 sg2 kernel: block drbd0: [drbd0_worker/2197]
sock_sendmsg time expired, ko = 4294967293
Oct 11 14:33:41 sg2 kernel: block drbd0: Out of sync: start=3932979480,
size=8 (sectors)
Oct 11 14:34:46 sg2 kernel: block drbd0: Out of sync: start=3946056120,
size=8 (sectors)
Oct 11 15:37:07 sg2 kernel: block drbd0: Out of sync: start=4696809024,
size=8 (sectors)
Oct 11 17:08:15 sg2 kernel: block drbd0: Out of sync: start=6084949528,
size=8 (sectors)
Oct 11 17:30:53 sg2 kernel: block drbd0: Out of sync: start=6567543472,
size=8 (sectors)
Oct 11 17:59:04 sg2 kernel: block drbd0: Out of sync: start=7169767896,
size=8 (sectors)
Oct 11 20:00:50 sg2 kernel: block drbd0: [drbd0_worker/2197]
sock_sendmsg time expired, ko = 4294967295
Oct 11 20:01:09 sg2 kernel: block drbd0: [drbd0_worker/2197]
sock_sendmsg time expired, ko = 4294967295
Oct 11 20:01:15 sg2 kernel: block drbd0: [drbd0_worker/2197]
sock_sendmsg time expired, ko = 4294967294
Oct 11 20:01:29 sg2 kernel: block drbd0: [drbd0_worker/2197]
sock_sendmsg time expired, ko = 4294967295
Oct 11 20:29:18 sg2 kernel: block drbd0: Out of sync: start=10362907296,
size=8 (sectors)
Oct 11 20:29:54 sg2 kernel: block drbd0: Out of sync: start=10375790488,
size=8 (sectors)
Oct 11 21:01:51 sg2 kernel: block drbd0: [drbd0_worker/2197]
sock_sendmsg time expired, ko = 4294967295
Oct 11 21:42:15 sg2 kernel: block drbd0: Out of sync: start=11907921096,
size=8 (sectors)
Oct 11 21:43:38 sg2 kernel: block drbd0: Out of sync: start=11937086248,
size=8 (sectors)
Oct 11 21:44:00 sg2 kernel: block drbd0: Out of sync: start=11944705032,
size=8 (sectors)
Oct 11 21:49:26 sg2 kernel: block drbd0: Out of sync: start=12062270432,
size=8 (sectors)
Oct 11 22:07:10 sg2 kernel: block drbd0: Out of sync: start=12440235128,
size=8 (sectors)
Oct 11 22:58:54 sg2 kernel: block drbd0: Out of sync: start=13548501984,
size=8 (sectors)
Oct 11 23:23:17 sg2 kernel: block drbd0: Out of sync: start=14072873320,
size=8 (sectors)
$ date
Sat Oct 11 23:28:11 PDT 2014

its 35% done at this point...   15 4K blocks out wrong of 1/3rd of 16TB
isn't a lot, but its still more than I like to see.

$ cat /proc/drbd
version: 8.3.15 (api:88/proto:86-97)
GIT-hash: 0ce4d235fc02b5c53c1c52c53433d11a694eab8c build by
phil@Build64R6, 2012-12-20 20:09:51
  0: cs:VerifyS ro:Secondary/Primary ds:UpToDate/UpToDate C r-----
     ns:0 nr:105707 dw:187685496 dr:654444832 al:0 bm:1 lo:107 pe:2104
ua:435 ap:0 ep:1 wo:f oos:60
         [=====>..............] verified: 34.6% (9846140/15051076)M
         finish: 14:55:27 speed: 187,648 (155,708) want: 204,800 K/sec




really, if I let this complete, then disconnect/reconnect the replica,
it will repair these glitches ?   I'm gathering I shoudl schedule these
verifies weekly or something.

That the backing device of one node fell out of sync is a cause concern. "Weekly" scan might be a bit much, but monthly or so isn't unreasonable. Of course, as you're seeing here, it's a lengthy process and it consumes non-trivial amounts of bandwidth and adds a fair load to the disks.

How long was it in production before this verify?

I can't speak to backuppc, but I am curious how you're managing the resources. Are you using cman + rgmanager or pacemaker?

--
Digimer
Papers and Projects: https://alteeve.ca/w/
What if the cure for cancer is trapped in the mind of a person without access to education?
_______________________________________________
CentOS mailing list
CentOS@xxxxxxxxxx
http://lists.centos.org/mailman/listinfo/centos




[Index of Archives]     [CentOS]     [CentOS Announce]     [CentOS Development]     [CentOS ARM Devel]     [CentOS Docs]     [CentOS Virtualization]     [Carrier Grade Linux]     [Linux Media]     [Asterisk]     [DCCP]     [Netdev]     [Xorg]     [Linux USB]
  Powered by Linux