> -----Original Message----- > From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf Of Alex Gorbachev > Sent: 23 August 2016 16:43 > To: Wido den Hollander <wido@xxxxxxxx> > Cc: ceph-users <ceph-users@xxxxxxxxxxxxxx>; Nick Fisk <nick@xxxxxxxxxx> > Subject: Re: udev rule to set readahead on Ceph RBD's > > On Mon, Aug 22, 2016 at 3:29 PM, Wido den Hollander <wido@xxxxxxxx> wrote: > > > >> Op 22 augustus 2016 om 21:22 schreef Nick Fisk <nick@xxxxxxxxxx>: > >> > >> > >> > -----Original Message----- > >> > From: Wido den Hollander [mailto:wido@xxxxxxxx] > >> > Sent: 22 August 2016 18:22 > >> > To: ceph-users <ceph-users@xxxxxxxxxxxxxx>; nick@xxxxxxxxxx > >> > Subject: Re: udev rule to set readahead on Ceph RBD's > >> > > >> > > >> > > Op 22 augustus 2016 om 15:17 schreef Nick Fisk <nick@xxxxxxxxxx>: > >> > > > >> > > > >> > > Hope it's useful to someone > >> > > > >> > > https://gist.github.com/fiskn/6c135ab218d35e8b53ec0148fca47bf6 > >> > > > >> > > >> > Thanks for sharing. Might this be worth adding it to ceph-common? > >> > >> Maybe, Ilya kindly set the default for krbd to 4MB last year in the kernel, but maybe having this available would be handy if people > ever want a different default. It could be set to 4MB as well, with a note somewhere to point people at its direction if they need to > change it. > >> > > > > I think it might be handy to have the udev file as redundancy. That way it can easily be changed by users. The udev file is already > present, they just have to modify it. > > > >> > > >> > And is 16MB something we should want by default or does this apply to your situation better? > >> > >> It sort of applies to me. With a 4MB readahead you will probably struggle to get much more than around 50-80MB/s sequential > reads, as the read ahead will only ever hit 1 object at a time. If you want to get nearer 200MB/s then you need to set either 16 or > 32MB readahead. I need it to stream to LTO6 tape. Depending on what you are doing this may or may not be required. > >> > > > > Ah, yes. I a kind of similar use-case I went for using 64MB objects underneath a RBD device. We needed high sequential Write and > Read performance on those RBD devices since we were storing large files on there. > > > > Different approach, kind of similar result. > > Question: what scheduler were you guys using to facilitate the readahead on the RBD client? Have you noticed any difference > between different elevators and have you tried blk-mq/scsi-mq? I thought since kernel 3.19 you didn't have a choice and RBD always used blk-mq? But that's what I'm using as default. > > Thank you. > -- > Alex Gorbachev > Storcium > > > > > > Wido > > > >> > > >> > Wido > >> > > >> > > > >> > > _______________________________________________ > >> > > ceph-users mailing list > >> > > ceph-users@xxxxxxxxxxxxxx > >> > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >> > > _______________________________________________ > > ceph-users mailing list > > ceph-users@xxxxxxxxxxxxxx > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com