Re: speedup ceph / scaling / find the bottleneck

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi, 
Stefan is on vacation for the moment,I don't know if he can reply you.

But I can reoly for him for the kvm part (as we do same tests together in parallel).

- kvm is 1.1
- rbd 0.48
- drive option rbd:pool/volume:auth_supported=cephx;none;keyring=/etc/pve/priv/ceph/ceph.keyring:mon_host=X.X.X.X";
-using writeback

writeback tuning in ceph.conf on the kvm host

rbd_cache_size = 33554432 
rbd_cache_max_age = 2.0 

benchmark use in kvm guest:
fio --filename=$DISK --direct=1 --rw=randwrite --bs=4k --size=200G --numjobs=50 --runtime=90 --group_reporting --name=file1

results show max 14000io/s with 1 vm, 7000io/s by vm with 2vm,...
so it doesn't scale

(bench is with directio, so maybe writeback cache don't help)

hardware for ceph , is 3 nodes with 4 intel ssd each. (1 drive can handle 40000io/s randwrite locally)


- Alexandre

----- Mail original ----- 

De: "Gregory Farnum" <greg@xxxxxxxxxxx> 
À: "Stefan Priebe" <s.priebe@xxxxxxxxxxxx> 
Cc: ceph-devel@xxxxxxxxxxxxxxx, "Sage Weil" <sage@xxxxxxxxxxx> 
Envoyé: Jeudi 5 Juillet 2012 23:33:18 
Objet: Re: speedup ceph / scaling / find the bottleneck 

Could you send over the ceph.conf on your KVM host, as well as how 
you're configuring KVM to use rbd? 

On Tue, Jul 3, 2012 at 11:20 AM, Stefan Priebe <s.priebe@xxxxxxxxxxxx> wrote: 
> I'm sorry but this is the KVM Host Machine there is no ceph running on this 
> machine. 
> 
> If i change the admin socket to: 
> admin_socket=/var/run/ceph_$name.sock 
> 
> i don't have any socket at all ;-( 
> 
> Am 03.07.2012 17:31, schrieb Sage Weil: 
> 
>> On Tue, 3 Jul 2012, Stefan Priebe - Profihost AG wrote: 
>>> 
>>> Hello, 
>>> 
>>> Am 02.07.2012 22:30, schrieb Josh Durgin: 
>>>> 
>>>> If you add admin_socket=/path/to/admin_socket for your client running 
>>>> qemu (in that client's ceph.conf section or manually in the qemu 
>>>> command line) you can check that caching is enabled: 
>>>> 
>>>> ceph --admin-daemon /path/to/admin_socket show config | grep rbd_cache 
>>>> 
>>>> And see statistics it generates (look for cache) with: 
>>>> 
>>>> ceph --admin-daemon /path/to/admin_socket perfcounters_dump 
>>> 
>>> 
>>> This doesn't work for me: 
>>> ceph --admin-daemon /var/run/ceph.sock show config 
>>> read only got 0 bytes of 4 expected for response length; invalid 
>>> command?2012-07-03 09:46:57.931821 7fa75d129700 -1 asok(0x8115a0) 
>>> AdminSocket: 
>>> request 'show config' not defined 
>> 
>> 
>> Oh, it's 'config show'. Also, 'help' will list the supported commands. 
>> 
>>> Also perfcounters does not show anything: 
>>> # ceph --admin-daemon /var/run/ceph.sock perfcounters_dump 
>>> {} 
>> 
>> 
>> There may be another daemon that tried to attach to the same socket file. 
>> You might want to set 'admin socket = /var/run/ceph/$name.sock' or 
>> something similar, or whatever else is necessary to make it a unique file. 
>> 
>>> ~]# ceph -v 
>>> ceph version 0.48argonaut-2-gb576faa 
>>> (commit:b576faa6f24356f4d3ec7205e298d58659e29c68) 
>> 
>> 
>> Out of curiousity, what patches are you applying on top of the release? 
>> 
>> sage 
>> 
> 
-- 
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in 
the body of a message to majordomo@xxxxxxxxxxxxxxx 
More majordomo info at http://vger.kernel.org/majordomo-info.html 



-- 

-- 



	

Alexandre D e rumier 

Ingénieur Systèmes et Réseaux 


Fixe : 03 20 68 88 85 

Fax : 03 20 68 90 88 


45 Bvd du Général Leclerc 59100 Roubaix 
12 rue Marivaux 75002 Paris 
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux