On Thu, Mar 08, 2018 at 07:06:25PM +0800, Ming Lei wrote: > On Thu, Mar 08, 2018 at 03:34:31PM +0530, Kashyap Desai wrote: > > > -----Original Message----- > > > From: Ming Lei [mailto:ming.lei@xxxxxxxxxx] > > > Sent: Thursday, March 8, 2018 6:46 AM > > > To: Kashyap Desai > > > Cc: Jens Axboe; linux-block@xxxxxxxxxxxxxxx; Christoph Hellwig; Mike > > Snitzer; > > > linux-scsi@xxxxxxxxxxxxxxx; Hannes Reinecke; Arun Easi; Omar Sandoval; > > > Martin K . Petersen; James Bottomley; Christoph Hellwig; Don Brace; > > Peter > > > Rivera; Laurence Oberman > > > Subject: Re: [PATCH V3 8/8] scsi: megaraid: improve scsi_mq performance > > via > > > .host_tagset > > > > > > On Wed, Mar 07, 2018 at 10:58:34PM +0530, Kashyap Desai wrote: > > > > > > > > > > > > Also one observation using V3 series patch. I am seeing below > > > > > > Affinity mapping whereas I have only 72 logical CPUs. It means we > > > > > > are really not going to use all reply queues. > > > > > > e.a If I bind fio jobs on CPU 18-20, I am seeing only one reply > > > > > > queue is used and that may lead to performance drop as well. > > > > > > > > > > If the mapping is in such shape, I guess it should be quite > > > > > difficult to > > > > figure out > > > > > one perfect way to solve this situation because one reply queue has > > > > > to > > > > handle > > > > > IOs submitted from 4~5 CPUs at average. > > > > > > > > 4.15.0-rc1 kernel has below mapping - I am not sure which commit id in > > " > > > > linux_4.16-rc-host-tags-v3.2" is changing the mapping of IRQ to CPU. > > > > It > > > > > > I guess the mapping you posted is read from /proc/irq/126/smp_affinity. > > > > > > If yes, no any patch in linux_4.16-rc-host-tags-v3.2 should change IRQ > > affinity > > > code, which is done in irq_create_affinity_masks(), as you saw, no any > > patch > > > in linux_4.16-rc-host-tags-v3.2 touches that code. > > > > > > Could you simply apply the patches in linux_4.16-rc-host-tags-v3.2 > > against > > > 4.15-rc1 kernel and see any difference? > > > > > > > will be really good if we can fall back to below mapping once again. > > > > Current repo linux_4.16-rc-host-tags-v3.2 is giving lots of random > > > > mapping of CPU - MSIx. And that will be problematic in performance > > run. > > > > > > > > As I posted earlier, latest repo will only allow us to use *18* reply > > > > > > Looks not see this report before, could you share us how you conclude > > that? > > > The only patch changing reply queue is the following one: > > > > > > https://marc.info/?l=linux-block&m=151972611911593&w=2 > > > > > > But not see any issue in this patch yet, can you recover to 72 reply > > queues > > > after reverting the patch in above link? > > Ming - > > > > While testing, my system went bad. I debug further and understood that > > affinity mapping was changed due to below commit - > > 84676c1f21e8ff54befe985f4f14dc1edc10046b > > > > [PATCH] genirq/affinity: assign vectors to all possible CPUs > > > > Because of above change, we end up using very less reply queue. Many reply > > queues on my setup was mapped to offline/not-available CPUs. This may be > > primary contributing to odd performance impact and it may not be truly due > > to V3/V4 patch series. > > Seems a good news, :-) > > > > > I am planning to check your V3 and V4 series after removing above commit > > ID (for performance impact.). > > You can run your test on a server in which all CPUs are kept as online > for avoiding this issue. > > Or you can apply the following patchset for avoiding this issue: > > https://marc.info/?l=linux-block&m=152050646332092&w=2 If you want to do this way, all patches have been put into the following tree(V4): https://github.com/ming1/linux/commits/v4.16-rc-host-tags-v4 #in reverse order genirq/affinity: irq vector spread among online CPUs as far as possible genirq/affinity: support to do irq vectors spread starting from any vector genirq/affinity: move actual irq vector spread into one helper genirq/affinity: rename *node_to_possible_cpumask as *node_to_cpumask scsi: megaraid: improve scsi_mq performance via .host_tagset scsi: hpsa: improve scsi_mq performance via .host_tagset block: null_blk: introduce module parameter of 'g_host_tags' scsi: Add template flag 'host_tagset' blk-mq: introduce BLK_MQ_F_HOST_TAGS blk-mq: introduce 'start_tag' field to 'struct blk_mq_tags' scsi: avoid to hold host_busy for scsi_mq scsi: read host_busy via scsi_host_busy() scsi: introduce scsi_host_busy() scsi: virtio_scsi: fix IO hang caused by irq vector automatic affinity scsi: introduce force_blk_mq scsi: megaraid_sas: fix selection of reply queue scsi: hpsa: fix selection of reply queue Thanks, Ming