Re: [RadosGW] Performance for Concurrency Connections

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hmm.... Interesting now. 

I have no admin socket opened around. 

root@p01:/var/run/ceph# ls /var/run/ceph -al
total 0
drwxr-xr-x  2 root root  40 Sep  9 07:47 .
drwxr-xr-x 17 root root 600 Sep 11 21:23 ..
root@p01:/var/run/ceph# lsof | grep radosgw.asok
root@p01:/var/run/ceph#

I review the on-line doc for radosgw : http://ceph.com/docs/next/radosgw/config-ref/
There's no configuration for rgw admin socket tho. 

root@s01:~# ceph --admin-daemon /var/run/ceph/ceph-osd.2.asok config show | grep rgw_thread
  "rgw_thread_pool_size": "100",

  1. I found that the OSD config information includes rgw_thread_pool_size , is this what you mentioned ?  
  2. Why that the value is on OSD?  
  3. Where is the value of rgw_thread_pool_size that OSDs referenced from ?  

+Hugo Kuo+
(+886) 935004793


2013/9/12 Yehuda Sadeh <yehuda@xxxxxxxxxxx>
On Wed, Sep 11, 2013 at 9:34 PM, Kuo Hugo <tonytkdk@xxxxxxxxx> wrote:
> Hi Yehuda,
>
> Here's my ceph.conf
>
> root@p01:/tmp# cat /etc/ceph/ceph.conf
> [global]
> fsid = 6e05675c-f545-4d88-9784-ea56ceda750e
> mon_initial_members = s01, s02, s03
> mon_host = 192.168.2.61,192.168.2.62,192.168.2.63
> auth_supported = cephx
> osd_journal_size = 1024
> filestore_xattr_use_omap = true
>
> [client.radosgw.gateway]
> host = p01
> keyring = /etc/ceph/keyring.radosgw.gateway
> rgw_socket_path = /tmp/radosgw.sock
> log_file = /var/log/ceph/radosgw.log
> rgw_thread_pool_size = 200
>
> Depends on my conf, the /tmp/radosgw.sock was created while starting radosgw
> service.
> So that I tried to show up config by :
>
> root@p01:/tmp# ceph --admin-daemon /tmp/radosgw.sock config show
> read only got 0 bytes of 4 expected for response length; invalid command?
>
> Is it a bug or operation mistake ?

You're connecting to the wrong socket. You need to connect to the
admin socket, not to the socket that used for web server <-> gateway
communication. That socket by default should reside in /var/run/ceph.


>
> root@p01:/tmp# radosgw-admin -v
> ceph version 0.61.8 (a6fdcca3bddbc9f177e4e2bf0d9cdd85006b028b)
>
>
> Appreciate ~
>
>
> +Hugo Kuo+
> (+886) 935004793
>
>
> 2013/9/11 Yehuda Sadeh <yehuda@xxxxxxxxxxx>
>>
>> On Wed, Sep 11, 2013 at 7:57 AM, Kuo Hugo <tonytkdk@xxxxxxxxx> wrote:
>> >
>> > Hi Yehuda,
>> >
>> > I tried ... a question for modifying param.
>> > How to make it effect to the RadosGW ?   is it by restarting radosgw ?
>> > The value was set to 200. I'm not sure if it's applied to RadosGW or
>> > not.
>> >
>> > Is there a way to check the runtime value of "rgw thread pool size" ?
>> >
>>
>> You can do it through the admin socket interface.
>> Try running something like:
>> $ ceph --admin-daemon /var/run/ceph/radosgw.asok config show
>>
>> $ ceph --admin-daemon /var/run/ceph/radosgw.asok config set
>> rgw_thread_pool_size 200
>>
>>
>> The path to the admin socket may be different, and in any case can be
>> set through the 'admin socket' variable in ceph.conf.
>>
>> Yehuda
>>
>>
>> >
>> >
>> > 2013/9/11 Yehuda Sadeh <yehuda@xxxxxxxxxxx>
>> >>
>> >> Try modifing the 'rgw thread pool size' param in your ceph.conf. By
>> >> default it's 100, so try increasing it and see if it affects anything.
>> >>
>> >> Yehuda
>> >>
>> >>
>> >> On Wed, Sep 11, 2013 at 3:14 AM, Kuo Hugo <tonytkdk@xxxxxxxxx> wrote:
>> >>>
>> >>> For ref :
>> >>>
>> >>> Benchmark result
>> >>>
>> >>> Could someone help me to improve the performance of high concurrency
>> >>> use case ?
>> >>>
>> >>> Any suggestion would be excellent.!
>> >>>
>> >>> +Hugo Kuo+
>> >>> (+886) 935004793
>> >>>
>
>

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux