On Fri, Jan 25, 2013 at 11:51 AM, Isaac Otsiabah <zmoo76b@xxxxxxxxx> wrote: > > > Gregory, the network physical layout is simple, the two networks are > separate. the 192.168.0 and the 192.168.1 are not subnets within a > network. Hi Isaac, Could you send us your routing tables on the osds (route -n). That's one more bit of information that might be useful for tracking this down. Thanks, -sam > > Isaac > > > > > ----- Original Message ----- > From: Gregory Farnum <greg@xxxxxxxxxxx> > To: Isaac Otsiabah <zmoo76b@xxxxxxxxx> > Cc: "ceph-devel@xxxxxxxxxxxxxxx" <ceph-devel@xxxxxxxxxxxxxxx> > Sent: Thursday, January 24, 2013 1:28 PM > Subject: Re: osd down (for 2 about 2 minutes) error after adding a new host to my cluster > > What's the physical layout of your networking? This additional log may prove helpful as well, but I really need a bit more context in evaluating the messages I see from the first one. :) > -Greg > > > On Thursday, January 24, 2013 at 9:24 AM, Isaac Otsiabah wrote: > >> >> >> Gregory, i tried send the the attached debug output several times and >> the mail server rejected them all probably becauseof the file size so i cut the log file size down and it is attached. You will see the >> reconnection failures by the error message line below. The ceph version >> is 0.56 >> >> >> it appears to be a timing issue because with the flag (debug ms=1) turned on, the system ran slower and became harder to fail. >> I >> ran it several times and finally got it to fail on (osd.0) using >> default crush map. The attached tar file contains log files for all >> components on g8ct plus the ceph.conf. By the way, the log file contain only the last 1384 lines where the error occurs. >> >> >> I started with a 1-node cluster on host g8ct (osd.0, osd.1, osd.2) and then added host g13ct (osd.3, osd.4, osd.5) >> >> >> id weight type name up/down reweight >> -1 6 root default >> -3 6 rack unknownrack >> -2 3 host g8ct >> 0 1 osd.0 down 1 >> 1 1 osd.1 up 1 >> 2 1 osd.2 up 1 >> -4 3 host g13ct >> 3 1 osd.3 up 1 >> 4 1 osd.4 up 1 >> 5 1 osd.5 up 1 >> >> >> >> The error messages are in ceph.log and ceph-osd.0.log: >> >> ceph.log:2013-01-08 >> 05:41:38.080470 osd.0 192.168.0.124:6801/25571 3 : [ERR] map e15 had >> wrong cluster addr (192.168.0.124:6802/25571 != my >> 192.168.1.124:6802/25571) >> ceph-osd.0.log:2013-01-08 05:41:38.080458 7f06757fa710 0 log [ERR] : map e15 had wrong cluster addr >> (192.168.0.124:6802/25571 != my 192.168.1.124:6802/25571) >> >> >> >> [root@g8ct ceph]# ceph -v >> ceph version 0.56 (1a32f0a0b42f169a7b55ed48ec3208f6d4edc1e8) >> >> >> Isaac >> >> >> ----- Original Message ----- >> From: Gregory Farnum <greg@xxxxxxxxxxx (mailto:greg@xxxxxxxxxxx)> >> To: Isaac Otsiabah <zmoo76b@xxxxxxxxx (mailto:zmoo76b@xxxxxxxxx)> >> Cc: "ceph-devel@xxxxxxxxxxxxxxx (mailto:ceph-devel@xxxxxxxxxxxxxxx)" <ceph-devel@xxxxxxxxxxxxxxx (mailto:ceph-devel@xxxxxxxxxxxxxxx)> >> Sent: Monday, January 7, 2013 1:27 PM >> Subject: Re: osd down (for 2 about 2 minutes) error after adding a new host to my cluster >> >> On Monday, January 7, 2013 at 1:00 PM, Isaac Otsiabah wrote: >> >> >> When i add a new host (with osd's) to my existing cluster, 1 or 2 >> previous osd(s) goes down for about 2 minutes and then they come back >> up. >> > >> > >> > [root@h1ct ~]# ceph osd tree >> > >> > # id weight type name up/down reweight >> > -1 >> > 3 root default >> > -3 3 rack unknownrack >> > -2 3 host h1 >> > 0 1 osd.0 up 1 >> > 1 1 osd.1 up 1 >> > 2 >> > 1 osd.2 up 1 >> >> >> For example, after adding host h2 (with 3 new osd) to the above cluster >> and running the "ceph osd tree" command, i see this: >> > >> > >> > [root@h1 ~]# ceph osd tree >> > >> > # id weight type name up/down reweight >> > -1 6 root default >> > -3 >> > 6 rack unknownrack >> > -2 3 host h1 >> > 0 1 osd.0 up 1 >> > 1 1 osd.1 down 1 >> > 2 >> > 1 osd.2 up 1 >> > -4 3 host h2 >> > 3 1 osd.3 up 1 >> > 4 1 osd.4 up >> > 1 >> > 5 1 osd.5 up 1 >> >> >> The down osd always come back up after 2 minutes or less andi see the >> following error message in the respective osd log file: >> > 2013-01-07 04:40:17.613028 7fec7f092760 1 journal _open >> > /ceph_journal/journals/journal_2 fd 26: 1073741824 bytes, block size >> > 4096 bytes, directio = 1, aio = 0 >> > 2013-01-07 04:40:17.613122 >> > 7fec7f092760 1 journal _open /ceph_journal/journals/journal_2 fd 26: >> > 1073741824 bytes, block size 4096 bytes, directio = 1, aio = 0 >> > 2013-01-07 >> > 04:42:10.006533 7fec746f7710 0 -- 192.168.0.124:6808/19449 >> >> > 192.168.1.123:6800/18287 pipe(0x7fec20000e10 sd=31 :6808 pgs=0 cs=0 >> > l=0).accept connect_seq 0 vs existing 0 state connecting >> > 2013-01-07 >> > 04:45:29.834341 7fec743f4710 0 -- 192.168.1.124:6808/19449 >> >> > 192.168.1.122:6800/20072 pipe(0x7fec5402f320 sd=28 :45438 pgs=7 cs=1 >> > l=0).fault, initiating reconnect >> > 2013-01-07 04:45:29.835748 >> > 7fec743f4710 0 -- 192.168.1.124:6808/19449 >> >> > 192.168.1.122:6800/20072 pipe(0x7fec5402f320 sd=28 :45439 pgs=15 cs=3 >> > l=0).fault, initiating reconnect >> > 2013-01-07 04:45:30.835219 7fec743f4710 0 -- >> > 192.168.1.124:6808/19449 >> 192.168.1.122:6800/20072 >> > pipe(0x7fec5402f320 sd=28 :45894 pgs=482 cs=903 l=0).fault, initiating >> > reconnect >> > 2013-01-07 04:45:30.837318 7fec743f4710 0 -- >> > 192.168.1.124:6808/19449 >> 192.168.1.122:6800/20072 >> > pipe(0x7fec5402f320 sd=28 :45895 pgs=483 cs=905 l=0).fault, initiating >> > reconnect >> > 2013-01-07 04:45:30.851984 7fec637fe710 0 log [ERR] : map >> > e27 had wrong cluster addr (192.168.0.124:6808/19449 != my >> > 192.168.1.124:6808/19449) >> > >> > Also, this only happens only when the cluster ip address and the public ip address are different for example >> > .... >> > .... >> > .... >> > [osd.0] >> > host = g8ct >> > public address = 192.168.0.124 >> > cluster address = 192.168.1.124 >> > btrfs devs = /dev/sdb >> > >> > .... >> > .... >> > >> > but does not happen when they are the same. Any idea what may be the issue? >> This isn't familiar to me at first glance. What version of Ceph are you using? >> >> If >> this is easy to reproduce, can you pastebin your ceph.conf and then add >> "debug ms = 1" to your global config and gather up the logs from each >> daemon? >> -Greg >> >> -- >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in >> the body of a message to majordomo@xxxxxxxxxxxxxxx (mailto:majordomo@xxxxxxxxxxxxxxx) >> More majordomo info at http://vger.kernel.org/majordomo >> >> >> Attachments: >> - ceph-osd.0.log.tar.gz >> > > > > -- > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html > > -- > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html