Re: understanding of multipathing and speed

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tuesday 06 July 2010 06:16:55 Bart Coninckx wrote:
> On Monday 05 July 2010 20:58:30 Christophe Varoqui wrote:
> > On lun., 2010-07-05 at 20:37 +0200, Bart Coninckx wrote:
> > > Hi,
> > >
> > > I would like to run my ideas by this list about multipathing and the
> > > results as far as storage speed is concerned.
> > >
> > > I'm using multipathing to two iSCSI targets pointing to the same
> > > storage. It was my understanding that this provides for network path
> > > redundancy (and it does, I tested this) but also for added speed.
> > > I did some tests with Bonnie++ however while both paths were active and
> > > one path was down and the results are basically the same.
> > >
> > > Am I assuming wrong things? Or have I configured things wrong?
> >
> > can you also include a 'multipath -l' output and sketch the
> > hba/switch/controller physical connections ?
> >
> > thanks,
> 
> Sure,
> 
> xen3:~ # multipath -l
> lx03 (1494554000000000000000000010000000000000002000000) dm-3
>  IET,VIRTUAL-DISK [size=10G][features=1 queue_if_no_path][hwhandler=0]
> \_ round-robin 0 [prio=-2][active]
>  \_ 2:0:0:0 sdc 8:32  [active][undef]
>  \_ 1:0:0:0 sdb 8:16  [active][undef]
> ws033 (1494554000000000000000000010000000100000002000000) dm-2 IET,VIRTUAL-
> DISK
> [size=15G][features=1 queue_if_no_path][hwhandler=0]
> \_ round-robin 0 [prio=-2][active]
>  \_ 2:0:0:1 sde 8:64  [active][undef]
>  \_ 1:0:0:1 sdd 8:48  [active][undef]
> ms01 (1494554000000000000000000010000000200000002000000) dm-1
>  IET,VIRTUAL-DISK [size=40G][features=1 queue_if_no_path][hwhandler=0]
> \_ round-robin 0 [prio=-2][active]
>  \_ 1:0:0:2 sdf 8:80  [active][undef]
>  \_ 2:0:0:2 sdg 8:96  [active][undef]
> 
> I have two Gigabit NICs in this server each running over a separate switch
>  to a separate gigabit NIC with a unique IP address on the storage IET
>  iSCSI target.
> 
> Is this sufficient info?
> 
> Thx,
> 
> Bart
> 
> --
> dm-devel mailing list
> dm-devel@xxxxxxxxxx
> https://www.redhat.com/mailman/listinfo/dm-devel
> 

Hi all,

to show my point, these are the results of running bonnie++ locally on the 
storage - the values I look at are Block values in K/sec in both sequential 
output (writing) and sequential input (reading):

Version 1.03e       ------Sequential Output------ --Sequential Input- --
Random-
                    -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --
Seeks--
Machine        Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP  /sec 
%CP
iscsi3           8G 69351  96 116112  32 41128  10 57874  82 107721  16 418.2   
0
                    ------Sequential Create------ --------Random 
Create--------
                    -Create-- --Read--- -Delete-- -Create-- --Read--- -
Delete--
              files  /sec %CP  /sec %CP  /sec %CP  /sec %CP  /sec %CP  /sec 
%CP
                 16  4533  99 +++++ +++ +++++ +++  4395  99 +++++ +++ 17122  
99
iscsi3,8G,69351,96,116112,32,41128,10,57874,82,107721,16,418.2,0,16,4533,99,
+++++,+++,+++++,+++,4395,99,+++++,+++,17122,99



So were are hitting roughly 110 MB/sec locally on the storage server.

Now these are the results do doing the same over multipath with two paths 
enabled:

Version 1.03e       ------Sequential Output------ --Sequential Input- --
Random-
                    -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --
Seeks--
Machine        Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP  /sec 
%CP
xen3             8G 63953  92 100525  26 26885   2 41957  55 68184   2 357.9   
0
                    ------Sequential Create------ --------Random 
Create--------
                    -Create-- --Read--- -Delete-- -Create-- --Read--- -
Delete--
              files  /sec %CP  /sec %CP  /sec %CP  /sec %CP  /sec %CP  /sec 
%CP
                 16  5326  98 +++++ +++ +++++ +++  5333  97 +++++ +++ 17179 
100
xen3,8G,63953,92,100525,26,26885,2,41957,55,68184,2,357.9,0,16,5326,98,+++++,
+++,+++++,+++,5333,97,+++++,+++,17179,100

You can see we hit somewhat less, probably due to TCP overhead (though this 
should cut things with 30%). Now the same with one path down:

Version 1.03e       ------Sequential Output------ --Sequential Input- --
Random-
                    -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --
Seeks--
Machine        Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP  /sec 
%CP
xen3             8G 33214  46 113811  29 27917   1 44474  58 68812   2 362.8   
0
                    ------Sequential Create------ --------Random 
Create--------
                    -Create-- --Read--- -Delete-- -Create-- --Read--- -
Delete--
              files  /sec %CP  /sec %CP  /sec %CP  /sec %CP  /sec %CP  /sec 
%CP
                 16  5294  98 +++++ +++ +++++ +++  5337  97 +++++ +++ 17183  
99
xen3,8G,33214,46,113811,29,27917,1,44474,58,68812,2,362.8,0,16,5294,98,+++++,
+++,+++++,+++,5337,97,+++++,+++,17183,99

As you can see, roughly the same K/sec for both output and input. Actually 
writing is even faster with one path down!
Can anyone make sense of these values?

thx!

B.
 

--
dm-devel mailing list
dm-devel@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/dm-devel


[Index of Archives]     [DM Crypt]     [Fedora Desktop]     [ATA RAID]     [Fedora Marketing]     [Fedora Packaging]     [Fedora SELinux]     [Yosemite Discussion]     [KDE Users]     [Fedora Docs]

  Powered by Linux