Re: High CPU consumption on Windows guest OS when libgfapi is used

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Vijay,

I finally manage to do this test on the shared volume.

gluster volume info

Volume Name: abctest
Type: Distributed-Replicate
Volume ID: 0db494e2-51a3-4521-a1ba-5d3479cecba2
Status: Started
Number of Bricks: 3 x 3 = 9
Transport-type: tcp
Bricks:
Brick1: abc11:/data/hdd1/abctest
Brick2: abc12:/data/hdd1/abctest
Brick3: abc14:/data/hdd1/abctest
Brick4: abc16:/data/hdd1/abctest
Brick5: abc17:/data/hdd1/abctest
Brick6: abc20:/data/hdd1/abctest
Brick7: abc22:/data/hdd1/abctest
Brick8: abc23:/data/hdd1/abctest
Brick9: abc24:/data/hdd1/abctest
Options Reconfigured:
features.shard-block-size: 16MB
features.shard: on
server.allow-insecure: on
storage.owner-uid: 165
storage.owner-gid: 165
nfs.disable: true
performance.quick-read: off
performance.io-cache: off
performance.read-ahead: off
performance.stat-prefetch: off
cluster.lookup-optimize: on
cluster.quorum-type: auto
cluster.server-quorum-type: server
transport.address-family: inet
performance.readdir-ahead: off

Result is still the same.

4k random write

IOPS5355.75
Avg. response time (ms)2.79
CPU utilization total (%)96.73
CPU Privilegde time (%)92.49

4k random read

4k random read
IOPS16718.93
Avg. response time (ms)0.9
CPU utilization total (%)79.2
CPU Privilegde time (%)75.43

The snapshot of top -H while running 4k random write

  PID USER      PR  NI    VIRT    RES    SHR S %CPU %MEM     TIME+ COMMAND
23850 qemu      20   0 10.294g 8.203g  11788 R 92.7  6.5   5:42.01 qemu-kvm
24116 qemu      20   0 10.294g 8.203g  11788 S 34.2  6.5   1:27.91 qemu-kvm
26948 qemu      20   0 10.294g 8.203g  11788 R 33.6  6.5   1:26.85 qemu-kvm
24115 qemu      20   0 10.294g 8.203g  11788 S 32.9  6.5   1:27.72 qemu-kvm
26937 qemu      20   0 10.294g 8.203g  11788 S 32.9  6.5   1:27.87 qemu-kvm
27050 qemu      20   0 10.294g 8.203g  11788 R 32.9  6.5   1:17.14 qemu-kvm
27033 qemu      20   0 10.294g 8.203g  11788 S 31.6  6.5   1:19.40 qemu-kvm
24119 qemu      20   0 10.294g 8.203g  11788 S 26.6  6.5   1:32.16 qemu-kvm
24120 qemu      20   0 10.294g 8.203g  11788 S 25.9  6.5   1:32.02 qemu-kvm
23880 qemu      20   0 10.294g 8.203g  11788 S  8.3  6.5   2:31.11 qemu-kvm
23881 qemu      20   0 10.294g 8.203g  11788 S  8.0  6.5   2:58.75 qemu-kvm
23878 qemu      20   0 10.294g 8.203g  11788 S  7.6  6.5   2:04.15 qemu-kvm
23879 qemu      20   0 10.294g 8.203g  11788 S  7.6  6.5   2:36.50 qemu-kvm

Thanks.

Cw

On Thu, Apr 21, 2016 at 10:12 PM, Vijay Bellur <vbellur@xxxxxxxxxx> wrote:
On Wed, Apr 20, 2016 at 4:17 AM, qingwei wei <tchengwee@xxxxxxxxx> wrote:
> Gluster volume configuration, those bold entries are the initial settings i
> have
>
> Volume Name: g37test
> Type: Stripe
> Volume ID: 3f9dae3d-08f9-4321-aeac-67f44c7eb1ac
> Status: Created
> Number of Bricks: 1 x 10 = 10
> Transport-type: tcp
> Bricks:
> Brick1: 192.168.123.4:/mnt/sdb_mssd/data
> Brick2: 192.168.123.4:/mnt/sdc_mssd/data
> Brick3: 192.168.123.4:/mnt/sdd_mssd/data
> Brick4: 192.168.123.4:/mnt/sde_mssd/data
> Brick5: 192.168.123.4:/mnt/sdf_mssd/data
> Brick6: 192.168.123.4:/mnt/sdg_mssd/data
> Brick7: 192.168.123.4:/mnt/sdh_mssd/data
> Brick8: 192.168.123.4:/mnt/sdj_mssd/data
> Brick9: 192.168.123.4:/mnt/sdm_mssd/data
> Brick10: 192.168.123.4:/mnt/sdn_mssd/data
> Options Reconfigured:
> server.allow-insecure: on
> storage.owner-uid: 165
> storage.owner-gid: 165
> performance.quick-read: off
> performance.io-cache: off
> performance.read-ahead: off
> performance.stat-prefetch: off
> cluster.eager-lock: enable
> network.remote-dio: enable
> cluster.quorum-type: auto
> cluster.server-quorum-type: server
> nfs.disable: true
>

I notice that you are using a stripe volume. Would it be possible to
test with a sharded volume? We will be focusing only on sharded
volumes for VM disks going forward.

Thanks,
Vijay

_______________________________________________
Gluster-devel mailing list
Gluster-devel@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-devel

[Index of Archives]     [Gluster Users]     [Ceph Users]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux