Re: bcache_gc: BUG: soft lockup

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, 11 May 2016, Jens-U. Mozdzen wrote:

> Hi *,
> 
> Zitat von Eric Wheeler <bcache@xxxxxxxxxxxxxxxxxx>:
> >On Mon, 2 May 2016, Yannis Aribaud wrote:
> >[...]
> >I think this is the first time I've heard of Ceph being used in a bcache
> >stack on the list.  Are there any others out there with success?  If so,
> >what kernel versions and disk stack configuration?
> 
> After an extensive test period, we have just started a productive Ceph
> environment on our bcache-based SAN servers:
> 
> - MD-RAID6 (several SAS disks) as bcache backing device
> - MD-RAID1 (two SAS SSDs) as bcache cache device, only for that single backing
> device
> - LVM on top of /dev/bcache0
> - LVs, xfs-formatted, mounted at a convenient place, used by OSDs

So no ceph here?  

If you're using 4.4.y then you definitely need the patch from Ming Lei.  
Read this (rather long) thread if you want the details:
  "block: make sure big bio is splitted into at most 256 bvecs"
This affects 4.3 and newer iirc.  

OTOH, 4.1 is rock solid.  As of 4.1.21 or so it has all of the bcache 
stability fixes to date.


> kernel on our SAN nodes is 4.1.13-5-default (64 bit), as distributed by
> OpenSUSE Leap 42.1 (SUSE makes sure vital bcache patches are included, amongst
> others).
> 
> We're planning to later switch to a similar setup like the OP is running,
> using separate disks with a common bcache caching device for OSDs.
> 
> While we have not stressed the Ceph part yet on the productive system (there's
> plenty of other data served by SCST/FC, NFS, SaMBa and others), we did not yet
> run into problems and especially no kernel crashes.
> 
> >[...]
> >Also, does your backing device(s) set raid_partial_stripes_expensive=1 in
> >queue_limits (eg, md raid5/6)?  I've seen bugs around that that might
> >not be fixed yet.
> 
> This does sound disturbing to me - could you please give more details,
> probably in a new thread?

Same as above I think.  When bcache writebacks in opt_io sized writes that 
exceed 256 bvecs then you run into issues.  It only does that if 
raid_partial_stripes_expensive=1 like raid5/6 when it tries to prevent 
re-writes to the same stride.  

--
Eric Wheeler


> 
> Regards,
> Jens
> 
--
To unsubscribe from this list: send the line "unsubscribe linux-bcache" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[Index of Archives]     [Linux ARM Kernel]     [Linux Filesystem Development]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [ECOS]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux