Re: Performance

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



hi Raghavendra G,

i use xen 3.4.1 on glusterfs. the images of virtual maschine are on glusterfs hosted. i run 16 domUs (xen images) on glusterfs an the disk speed is awful slow. and in monitoring tool i see that there is a lot of wait cpu (that means that the cpus of the domUs are waiting on glusterfs). before i switched on glusterfs the domU images laying on local disk and there are no wait cpu....

best regards roland

Am 18.03.2010 05:46, schrieb Raghavendra G:
Hi Roland,

What are the applications you are using on glusterfs? In particular, what is the i/o pattern of applications? As a general guideline, you can try enabling/disabling each of the performance translators and observe the gain/loss of performance and tune the configuration accordingly.

regards,
On Wed, Mar 17, 2010 at 9:05 PM, Roland Fischer <roland.fischer@xxxxxxxxxx <mailto:roland.fischer@xxxxxxxxxx>> wrote:

    Hi Community,

    i need your help. i have performance problems with glusterfs 3.0.0
    and domUs (xen)

    i use 2 ident glusterfs-server (physikal HW) and two xen-server
    (physikal)

    currently i use client side replication - which is awful slow. i
    use a monitoring tool and see in domUs that there is a lot of cpu
    waiting. (before i switch to glusterfs there was no wait CPU)

    is server-side-replication faster and failsave. i mean if one
    glusterfs server goes down, does the other take over the domUs?

    is there anything in volfiles which i can tune?!? should i use
    server-side-replication?!?

    should i use the --disable-direct-io-mode? if yes on server side
    or client-side or both  -  and how to add in fstab (with
    --disable-direct-io-mode)?????

    Thank you for your help!!!

    servervolfile:
    cat /etc/glusterfs/export-domU-images-client_repl.vol
    #############
    volume posix
     type storage/posix
     option directory /GFS/domU-images
    end-volume

    volume locks
     type features/locks
     subvolumes posix
    end-volume

    volume domU-images
     type performance/io-threads
     option thread-count 8 # default is 16
     subvolumes locks
    end-volume

    volume server
     type protocol/server
     option transport-type tcp
     option auth.addr.domU-images.allow 192.*.*.*,127.0.0.1
     option transport.socket.listen-port 6997
     subvolumes domU-images
    end-volume
    ######################

    clientvolfiles:

    cat /etc/glusterfs/mount-domU-images-client_repl.vol
    volume gfs-01-01
     type protocol/client
     option transport-type tcp
     option remote-host hostname
     option transport.socket.nodelay on
     option remote-port 6997
     option remote-subvolume domU-images
     option ping-timeout 5
    end-volume

    volume gfs-01-02
     type protocol/client
     option transport-type tcp
     option remote-host hostname
     option transport.socket.nodelay on
     option remote-port 6997
     option remote-subvolume domU-images
     option ping-timeout 5
    end-volume

    volume gfs-replicate
     type cluster/replicate
     subvolumes gfs-01-01 gfs-01-02
    end-volume

    volume writebehind
     type performance/write-behind
     option cache-size 4MB   #default 16
     subvolumes gfs-replicate
    end-volume

    volume readahead
     type performance/read-ahead
     option page-count 8              # cache per file  = (page-count
    x page-size)
     subvolumes writebehind
    end-volume

    volume iocache
     type performance/io-cache
     option cache-size 1GB   #new 1GB supported
     option cache-timeout 1
     subvolumes readahead
    end-volume

    volume statprefetch
       type performance/stat-prefetch
       subvolumes iocache
    end-volume

    #################################################


    Best regards,
    Roland



    _______________________________________________
    Gluster-devel mailing list
    Gluster-devel@xxxxxxxxxx <mailto:Gluster-devel@xxxxxxxxxx>
    http://lists.nongnu.org/mailman/listinfo/gluster-devel




--
Raghavendra G






[Index of Archives]     [Gluster Users]     [Ceph Users]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux