Re: Slow write times to gluster disk

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



----- Original Message -----
> From: "Pranith Kumar Karampuri" <pkarampu@xxxxxxxxxx>
> To: "Pat Haley" <phaley@xxxxxxx>
> Cc: gluster-users@xxxxxxxxxxx, "Steve Postma" <SPostma@xxxxxxxxxxxx>
> Sent: Friday, May 12, 2017 11:17:11 PM
> Subject: Re:  Slow write times to gluster disk
> 
> 
> 
> On Sat, May 13, 2017 at 8:44 AM, Pranith Kumar Karampuri <
> pkarampu@xxxxxxxxxx > wrote:
> 
> 
> 
> 
> 
> On Fri, May 12, 2017 at 8:04 PM, Pat Haley < phaley@xxxxxxx > wrote:
> 
> 
> 
> 
> Hi Pranith,
> 
> My question was about setting up a gluster volume on an ext4 partition. I
> thought we had the bricks mounted as xfs for compatibility with gluster?
> 
> Oh that should not be a problem. It works fine.
> 
> Just that xfs doesn't have limits for anything, where as ext4 does for things
> like hardlinks etc(At least last time I checked :-) ). So it is better to
> have xfs.

One of the biggest reasons to use XFS IMHO is that most of the testing / large scale deployments(at least that I know of) / etc are done using XFS as a backend.  While EXT4 should work I don't think that it has the same level of testing as XFS.

-b 



> 
> 
> 
> 
> 
> 
> 
> Pat
> 
> 
> 
> On 05/11/2017 12:06 PM, Pranith Kumar Karampuri wrote:
> 
> 
> 
> 
> 
> On Thu, May 11, 2017 at 9:32 PM, Pat Haley < phaley@xxxxxxx > wrote:
> 
> 
> 
> 
> Hi Pranith,
> 
> The /home partition is mounted as ext4
> /home ext4 defaults,usrquota,grpquota 1 2
> 
> The brick partitions are mounted ax xfs
> /mnt/brick1 xfs defaults 0 0
> /mnt/brick2 xfs defaults 0 0
> 
> Will this cause a problem with creating a volume under /home?
> 
> I don't think the bottleneck is disk. You can do the same tests you did on
> your new volume to confirm?
> 
> 
> 
> 
> Pat
> 
> 
> 
> On 05/11/2017 11:32 AM, Pranith Kumar Karampuri wrote:
> 
> 
> 
> 
> 
> On Thu, May 11, 2017 at 8:57 PM, Pat Haley < phaley@xxxxxxx > wrote:
> 
> 
> 
> 
> Hi Pranith,
> 
> Unfortunately, we don't have similar hardware for a small scale test. All we
> have is our production hardware.
> 
> You said something about /home partition which has lesser disks, we can
> create plain distribute volume inside one of those directories. After we are
> done, we can remove the setup. What do you say?
> 
> 
> 
> 
> 
> Pat
> 
> 
> 
> 
> On 05/11/2017 07:05 AM, Pranith Kumar Karampuri wrote:
> 
> 
> 
> 
> 
> On Thu, May 11, 2017 at 2:48 AM, Pat Haley < phaley@xxxxxxx > wrote:
> 
> 
> 
> 
> Hi Pranith,
> 
> Since we are mounting the partitions as the bricks, I tried the dd test
> writing to <brick-path>/.glusterfs/<file-to-be-removed-after-test>. The
> results without oflag=sync were 1.6 Gb/s (faster than gluster but not as
> fast as I was expecting given the 1.2 Gb/s to the no-gluster area w/ fewer
> disks).
> 
> Okay, then 1.6Gb/s is what we need to target for, considering your volume is
> just distribute. Is there any way you can do tests on similar hardware but
> at a small scale? Just so we can run the workload to learn more about the
> bottlenecks in the system? We can probably try to get the speed to 1.2Gb/s
> on your /home partition you were telling me yesterday. Let me know if that
> is something you are okay to do.
> 
> 
> 
> 
> Pat
> 
> 
> 
> On 05/10/2017 01:27 PM, Pranith Kumar Karampuri wrote:
> 
> 
> 
> 
> 
> On Wed, May 10, 2017 at 10:15 PM, Pat Haley < phaley@xxxxxxx > wrote:
> 
> 
> 
> 
> Hi Pranith,
> 
> Not entirely sure (this isn't my area of expertise). I'll run your answer by
> some other people who are more familiar with this.
> 
> I am also uncertain about how to interpret the results when we also add the
> dd tests writing to the /home area (no gluster, still on the same machine)
> 
> 
>     * dd test without oflag=sync (rough average of multiple tests)
> 
> 
>         * gluster w/ fuse mount : 570 Mb/s
>         * gluster w/ nfs mount: 390 Mb/s
>         * nfs (no gluster): 1.2 Gb/s
>     * dd test with oflag=sync (rough average of multiple tests)
> 
>         * gluster w/ fuse mount: 5 Mb/s
>         * gluster w/ nfs mount: 200 Mb/s
>         * nfs (no gluster): 20 Mb/s
> 
> Given that the non-gluster area is a RAID-6 of 4 disks while each brick of
> the gluster area is a RAID-6 of 32 disks, I would naively expect the writes
> to the gluster area to be roughly 8x faster than to the non-gluster.
> 
> I think a better test is to try and write to a file using nfs without any
> gluster to a location that is not inside the brick but someother location
> that is on same disk(s). If you are mounting the partition as the brick,
> then we can write to a file inside .glusterfs directory, something like
> <brick-path>/.glusterfs/<file-to-be-removed-after-test>.
> 
> 
> 
> 
> 
> I still think we have a speed issue, I can't tell if fuse vs nfs is part of
> the problem.
> 
> I got interested in the post because I read that fuse speed is lesser than
> nfs speed which is counter-intuitive to my understanding. So wanted
> clarifications. Now that I got my clarifications where fuse outperformed nfs
> without sync, we can resume testing as described above and try to find what
> it is. Based on your email-id I am guessing you are from Boston and I am
> from Bangalore so if you are okay with doing this debugging for multiple
> days because of timezones, I will be happy to help. Please be a bit patient
> with me, I am under a release crunch but I am very curious with the problem
> you posted.
> 
> 
> 
> 
> Was there anything useful in the profiles?
> 
> Unfortunately profiles didn't help me much, I think we are collecting the
> profiles from an active volume, so it has a lot of information that is not
> pertaining to dd so it is difficult to find the contributions of dd. So I
> went through your post again and found something I didn't pay much attention
> to earlier i.e. oflag=sync, so did my own tests on my setup with FUSE so
> sent that reply.
> 
> 
> 
> 
> 
> Pat
> 
> 
> 
> On 05/10/2017 12:15 PM, Pranith Kumar Karampuri wrote:
> 
> 
> 
> Okay good. At least this validates my doubts. Handling O_SYNC in gluster NFS
> and fuse is a bit different.
> When application opens a file with O_SYNC on fuse mount then each write
> syscall has to be written to disk as part of the syscall where as in case of
> NFS, there is no concept of open. NFS performs write though a handle saying
> it needs to be a synchronous write, so write() syscall is performed first
> then it performs fsync(). so an write on an fd with O_SYNC becomes
> write+fsync. I am suspecting that when multiple threads do this
> write+fsync() operation on the same file, multiple writes are batched
> together to be written do disk so the throughput on the disk is increasing
> is my guess.
> 
> Does it answer your doubts?
> 
> On Wed, May 10, 2017 at 9:35 PM, Pat Haley < phaley@xxxxxxx > wrote:
> 
> 
> 
> 
> Without the oflag=sync and only a single test of each, the FUSE is going
> faster than NFS:
> 
> FUSE:
> mseas-data2(dri_nascar)% dd if=/dev/zero count=4096 bs=1048576 of=zeros.txt
> conv=sync
> 4096+0 records in
> 4096+0 records out
> 4294967296 bytes (4.3 GB) copied, 7.46961 s, 575 MB/s
> 
> 
> NFS
> mseas-data2(HYCOM)% dd if=/dev/zero count=4096 bs=1048576 of=zeros.txt
> conv=sync
> 4096+0 records in
> 4096+0 records out
> 4294967296 bytes (4.3 GB) copied, 11.4264 s, 376 MB/s
> 
> 
> 
> On 05/10/2017 11:53 AM, Pranith Kumar Karampuri wrote:
> 
> 
> 
> Could you let me know the speed without oflag=sync on both the mounts? No
> need to collect profiles.
> 
> On Wed, May 10, 2017 at 9:17 PM, Pat Haley < phaley@xxxxxxx > wrote:
> 
> 
> 
> 
> Here is what I see now:
> 
> [root@mseas-data2 ~]# gluster volume info
> 
> Volume Name: data-volume
> Type: Distribute
> Volume ID: c162161e-2a2d-4dac-b015-f31fd89ceb18
> Status: Started
> Number of Bricks: 2
> Transport-type: tcp
> Bricks:
> Brick1: mseas-data2:/mnt/brick1
> Brick2: mseas-data2:/mnt/brick2
> Options Reconfigured:
> diagnostics.count-fop-hits: on
> diagnostics.latency-measurement: on
> nfs.exports-auth-enable: on
> diagnostics.brick-sys-log-level: WARNING
> performance.readdir-ahead: on
> nfs.disable: on
> nfs.export-volumes: off
> 
> 
> 
> On 05/10/2017 11:44 AM, Pranith Kumar Karampuri wrote:
> 
> 
> 
> Is this the volume info you have?
> 
> > [ root at mseas-data2 ~]# gluster volume info > > Volume Name: data-volume
> > > Type: Distribute > Volume ID: c162161e-2a2d-4dac-b015-f31fd89ceb18 >
> > Status: Started > Number of Bricks: 2 > Transport-type: tcp > Bricks: >
> > Brick1: mseas-data2:/mnt/brick1 > Brick2: mseas-data2:/mnt/brick2 >
> > Options Reconfigured: > performance.readdir-ahead: on > nfs.disable: on >
> > nfs.export-volumes: off
> ​I copied this from old thread from 2016. This is distribute volume. Did you
> change any of the options in between?
> --
> 
> -=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
> Pat Haley                          Email: phaley@xxxxxxx Center for Ocean
> Engineering       Phone:  (617) 253-6824
> Dept. of Mechanical Engineering    Fax:    (617) 253-8125
> MIT, Room 5-213 http://web.mit.edu/phaley/www/ 77 Massachusetts Avenue
> Cambridge, MA  02139-4301
> --
> Pranith
> --
> 
> -=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
> Pat Haley                          Email: phaley@xxxxxxx Center for Ocean
> Engineering       Phone:  (617) 253-6824
> Dept. of Mechanical Engineering    Fax:    (617) 253-8125
> MIT, Room 5-213 http://web.mit.edu/phaley/www/ 77 Massachusetts Avenue
> Cambridge, MA  02139-4301
> --
> Pranith
> --
> 
> -=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
> Pat Haley                          Email: phaley@xxxxxxx Center for Ocean
> Engineering       Phone:  (617) 253-6824
> Dept. of Mechanical Engineering    Fax:    (617) 253-8125
> MIT, Room 5-213 http://web.mit.edu/phaley/www/ 77 Massachusetts Avenue
> Cambridge, MA  02139-4301
> --
> Pranith
> --
> 
> -=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
> Pat Haley                          Email: phaley@xxxxxxx Center for Ocean
> Engineering       Phone:  (617) 253-6824
> Dept. of Mechanical Engineering    Fax:    (617) 253-8125
> MIT, Room 5-213 http://web.mit.edu/phaley/www/ 77 Massachusetts Avenue
> Cambridge, MA  02139-4301
> --
> Pranith
> --
> 
> -=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
> Pat Haley                          Email: phaley@xxxxxxx Center for Ocean
> Engineering       Phone:  (617) 253-6824
> Dept. of Mechanical Engineering    Fax:    (617) 253-8125
> MIT, Room 5-213 http://web.mit.edu/phaley/www/ 77 Massachusetts Avenue
> Cambridge, MA  02139-4301
> --
> Pranith
> --
> 
> -=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
> Pat Haley                          Email: phaley@xxxxxxx Center for Ocean
> Engineering       Phone:  (617) 253-6824
> Dept. of Mechanical Engineering    Fax:    (617) 253-8125
> MIT, Room 5-213 http://web.mit.edu/phaley/www/ 77 Massachusetts Avenue
> Cambridge, MA  02139-4301
> --
> Pranith
> --
> 
> -=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
> Pat Haley                          Email: phaley@xxxxxxx Center for Ocean
> Engineering       Phone:  (617) 253-6824
> Dept. of Mechanical Engineering    Fax:    (617) 253-8125
> MIT, Room 5-213 http://web.mit.edu/phaley/www/ 77 Massachusetts Avenue
> Cambridge, MA  02139-4301
> 
> 
> 
> --
> Pranith
> 
> 
> 
> --
> Pranith
> 
> _______________________________________________
> Gluster-users mailing list
> Gluster-users@xxxxxxxxxxx
> http://lists.gluster.org/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://lists.gluster.org/mailman/listinfo/gluster-users




[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux