> -----Original Message----- > From: ceph-devel-owner@xxxxxxxxxxxxxxx [mailto:ceph-devel- > owner@xxxxxxxxxxxxxxx] On Behalf Of Sage Weil > Sent: Thursday, September 24, 2015 3:32 AM > To: Handzik, Joe > Cc: Somnath Roy; Samuel Just; Samuel Just (sam.just@xxxxxxxxxxx); ceph- > devel > Subject: Re: Very slow recovery/peering with latest master > > On Wed, 23 Sep 2015, Handzik, Joe wrote: > > Ok. When configuring with ceph-disk, it does something nifty and > > actually gives the OSD the uuid of the disk's partition as its fsid. I > > bootstrap off that to get an argument to pass into the function you > > have identified as the bottleneck. I ran it by sage and we both > > realized there would be cases where it wouldn't work...I'm sure > > neither of us realized the failure would take three minutes though. > > > > In the short term, it makes sense to create an option to disable or > > short-circuit the blkid code. I would prefer that the default be left > > with the code enabled, but I'm open to default disabled if others > > think this will be a widespread problem. You could also make sure your > > OSD fsids are set to match your disk partition uuids for now too, if > > that's a faster workaround for you (it'll get rid of the failure). > > I think we should try to figure out where it is hanging. Can you strace the > blkid process to see what it is up to? > > I opened http://tracker.ceph.com/issues/13219 > > I think as long as it behaves reliably with ceph-disk OSDs then we can have it > on by default. > > sage > > > > > > Joe > > > > > On Sep 23, 2015, at 6:26 PM, Somnath Roy <Somnath.Roy@xxxxxxxxxxx> > wrote: > > > > > > <<inline > > > > > > -----Original Message----- > > > From: Handzik, Joe [mailto:joseph.t.handzik@xxxxxxx] > > > Sent: Wednesday, September 23, 2015 4:20 PM > > > To: Samuel Just > > > Cc: Somnath Roy; Samuel Just (sam.just@xxxxxxxxxxx); Sage Weil > > > (sage@xxxxxxxxxxxx); ceph-devel > > > Subject: Re: Very slow recovery/peering with latest master > > > > > > I added that, there is code up the stack in calamari that consumes the > path provided, which is intended in the future to facilitate disk monitoring > and management. > > > > > > [Somnath] Ok > > > > > > Somnath, what does your disk configuration look like (filesystem, > SSD/HDD, anything else you think could be relevant)? Did you configure your > disks with ceph-disk, or by hand? I never saw this while testing my code, has > anyone else heard of this behavior on master? The code has been in master > for 2-3 months now I believe. > > > [Somnath] All SSD , I use mkcephfs to create cluster , I partitioned the > disk with fdisk beforehand. I am using XFS. Are you trying with Ubuntu 3.16.* > kernel ? It could be Linux distribution/kernel specific. Somnath, maybe it is GPT related, what partition table do you have? I think parted and gdisk can create GPT partitions, but not fdisk (definitely not in version that I use). You could backup and clear blkid cache /etc/blkid/blkid.tab, maybe there is a mess. Regards, Igor. > > > > > > It would be nice to not need to disable this, but if this behavior exists and > can't be explained by a misconfiguration or something else I'll need to figure > out a different implementation. > > > > > > Joe > > > > > >> On Sep 23, 2015, at 6:07 PM, Samuel Just <sjust@xxxxxxxxxx> wrote: > > >> > > >> Wow. Why would that take so long? I think you are correct that > > >> it's only used for metadata, we could just add a config value to > > >> disable it. > > >> -Sam > > >> > > >>> On Wed, Sep 23, 2015 at 3:48 PM, Somnath Roy > <Somnath.Roy@xxxxxxxxxxx> wrote: > > >>> Sam/Sage, > > >>> I debugged it down and found out that the get_device_by_uuid- > >blkid_find_dev_with_tag() call within FileStore::collect_metadata() is > hanging for ~3 mins before returning a EINVAL. I saw this portion is newly > added after hammer. > > >>> Commenting it out resolves the issue. BTW, I saw this value is stored as > metadata but not used anywhere , am I missing anything ? > > >>> Here is my Linux details.. > > >>> > > >>> root@emsnode5:~/wip-write-path-optimization/src# uname -a Linux > > >>> emsnode5 3.16.0-38-generic #52~14.04.1-Ubuntu SMP Fri May 8 > > >>> 09:43:57 UTC 2015 x86_64 x86_64 x86_64 GNU/Linux > > >>> > > >>> > > >>> root@emsnode5:~/wip-write-path-optimization/src# lsb_release -a > No > > >>> LSB modules are available. > > >>> Distributor ID: Ubuntu > > >>> Description: Ubuntu 14.04.2 LTS > > >>> Release: 14.04 > > >>> Codename: trusty > > >>> > > >>> Thanks & Regards > > >>> Somnath > > >>> > > >>> -----Original Message----- > > >>> From: Somnath Roy > > >>> Sent: Wednesday, September 16, 2015 2:20 PM > > >>> To: 'Gregory Farnum' > > >>> Cc: 'ceph-devel' > > >>> Subject: RE: Very slow recovery/peering with latest master > > >>> > > >>> > > >>> Sage/Greg, > > >>> > > >>> Yeah, as we expected, it is not happening probably because of > recovery settings. I reverted it back in my ceph.conf , but, still seeing this > problem. > > >>> > > >>> Some observation : > > >>> ---------------------- > > >>> > > >>> 1. First of all, I don't think it is something related to my environment. I > recreated the cluster with Hammer and this problem is not there. > > >>> > > >>> 2. I have enabled the messenger/monclient log (Couldn't attach here) > in one of the OSDs and found monitor is taking long time to detect the up > OSDs. If you see the log, I have started OSD at 2015-09-16 16:13:07.042463 , > but, there is no communication (only getting KEEP_ALIVE) till 2015-09-16 > 16:16:07.180482 , so, 3 mins !! > > >>> > > >>> 3. During this period, I saw monclient trying to communicate with > monitor but not able to probably. It is sending osd_boot at 2015-09-16 > 16:16:07.180482 only.. > > >>> > > >>> 2015-09-16 16:16:07.180450 7f65377fe700 10 monclient: > > >>> _send_mon_message to mon.a at 10.60.194.10:6789/0 > > >>> 2015-09-16 16:16:07.180482 7f65377fe700 1 -- > > >>> 10.60.194.10:6820/20102 > > >>> --> 10.60.194.10:6789/0 -- osd_boot(osd.10 booted 0 features > > >>> 72057594037927935 v45) v6 -- ?+0 0x7f6523c19100 con 0x7f6542045680 > > >>> 2015-09-16 16:16:07.180496 7f65377fe700 20 -- 10.60.194.10:6820/20102 > submit_message osd_boot(osd.10 booted 0 features 72057594037927935 > v45) v6 remote, 10.60.194.10:6789/0, have pipe. > > >>> > > >>> 4. BTW, the osd down scenario is detected very quickly (ceph -w > output) , problem is during coming up I guess. > > >>> > > >>> > > >>> So, something related to mon communication getting slower ? > > >>> Let me know if more verbose logging is required and how should I > share the log.. > > >>> > > >>> Thanks & Regards > > >>> Somnath > > >>> > > >>> -----Original Message----- > > >>> From: Gregory Farnum [mailto:gfarnum@xxxxxxxxxx] > > >>> Sent: Wednesday, September 16, 2015 11:35 AM > > >>> To: Somnath Roy > > >>> Cc: ceph-devel > > >>> Subject: Re: Very slow recovery/peering with latest master > > >>> > > >>>> On Tue, Sep 15, 2015 at 8:04 PM, Somnath Roy > <Somnath.Roy@xxxxxxxxxxx> wrote: > > >>>> Hi, > > >>>> I am seeing very slow recovery when I am adding OSDs with the latest > master. > > >>>> Also, If I just restart all the OSDs (no IO is going on in the cluster) , > cluster is taking a significant amount of time to reach in active+clean state > (and even detecting all the up OSDs). > > >>>> > > >>>> I saw the recovery/backfill default parameters are now changed (to > lower value) , this probably explains the recovery scenario , but, will it affect > the peering time during OSD startup as well ? > > >>> > > >>> I don't think these values should impact peering time, but you could > configure them back to the old defaults and see if it changes. > > >>> -Greg > > >>> > > >>> ________________________________ > > >>> > > >>> PLEASE NOTE: The information contained in this electronic mail > message is intended only for the use of the designated recipient(s) named > above. If the reader of this message is not the intended recipient, you are > hereby notified that you have received this message in error and that any > review, dissemination, distribution, or copying of this message is strictly > prohibited. If you have received this communication in error, please notify > the sender by telephone or e-mail (as shown above) immediately and > destroy any and all copies of this message in your possession (whether hard > copies or electronically stored copies). > > >> -- > > >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" > > >> in the body of a message to majordomo@xxxxxxxxxxxxxxx More > > >> majordomo info at http://vger.kernel.org/majordomo-info.html > > -- > > To unsubscribe from this list: send the line "unsubscribe ceph-devel" > > in the body of a message to majordomo@xxxxxxxxxxxxxxx More > majordomo > > info at http://vger.kernel.org/majordomo-info.html > > > > > -- > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the > body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at > http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html