radosgw warnings in apache error.log

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



We are testing radosgw with cyberduck, so far we see the following issues

1. In apache error log for each file put we see:

[Tue Sep 03 17:35:24 2013] [warn] FastCGI: 193.218.104.138 PUT https://193.218.100.131/test/tesfile04.iso auth AWS ***
[Tue Sep 03 17:35:24 2013] [warn] FastCGI: JJJ gotCont=1

- This is with the ceph version of apache/fastcgi.
- This happens on each file, 1 time at the beginning of the transfer and at least at large files, several times at the end. 
- The standard apache/fastcgi combo does not show this issue.
About what do we get warned here?


2. Large file transfer of 4GB iso image with cyberduck
Uploads of a 4GB .iso file fail at the end of each upload, the file is then shown with 0B in directory listing

At the end of the upload I get the following in rados log
2013-09-03 18:00:49.427092 7faf87fff700  2 RGWDataChangesLog::ChangesRenewThread: start
2013-09-03 18:01:11.427201 7faf87fff700  2 RGWDataChangesLog::ChangesRenewThread: start
2013-09-03 18:01:33.427310 7faf87fff700  2 RGWDataChangesLog::ChangesRenewThread: start
2013-09-03 18:01:36.557121 7faf46ff5700 10 x>> x-amz-acl:public-read
2013-09-03 18:01:36.557567 7faf46ff5700 20 get_obj_state: rctx=0x7faf540251c0 obj=test-han:tesfile05.iso state=0x7faf5412ecd8 s->prefetch_data=0
2013-09-03 18:01:36.558581 7faf46ff5700  0 setting object write_tag=default.6641.36
2013-09-03 18:01:36.574302 7faf46ff5700 20 get_obj_state: rctx=0x7faf540251c0 obj=test-han:_shadow__Bw6LkKRYSmAwEh5VBmxo7fb5Md-IY7d_1 state=0x7faf54028488 s->prefetch_data=0
2013-09-03 18:01:36.575215 7faf46ff5700 20 get_obj_state: s->obj_tag was set empty
2013-09-03 18:01:36.575250 7faf46ff5700 20 prepare_atomic_for_write_impl: state is not atomic. state=0x7faf54028488
2013-09-03 18:01:36.579593 7faf46ff5700 20 get_obj_state: rctx=0x7faf540251c0 obj=test-han:_shadow__Bw6LkKRYSmAwEh5VBmxo7fb5Md-IY7d_2 state=0x7faf5412ea28 s->prefetch_data=0
2013-09-03 18:01:36.580581 7faf46ff5700 20 get_obj_state: s->obj_tag was set empty
2013-09-03 18:01:36.580635 7faf46ff5700 20 prepare_atomic_for_write_impl: state is not atomic. state=0x7faf5412ea28
2013-09-03 18:01:36.585058 7faf46ff5700 20 get_obj_state: rctx=0x7faf540251c0 obj=test-han:_shadow__Bw6LkKRYSmAwEh5VBmxo7fb5Md-IY7d_3 state=0x7faf5412aaa8 s->prefetch_data=0
2013-09-03 18:01:36.585956 7faf46ff5700 20 get_obj_state: s->obj_tag was set empty
2013-09-03 18:01:36.586005 7faf46ff5700 20 prepare_atomic_for_write_impl: state is not atomic. state=0x7faf5412aaa8
2013-09-03 18:01:36.589811 7faf46ff5700 20 get_obj_state: rctx=0x7faf540251c0 obj=test-han:_shadow__Bw6LkKRYSmAwEh5VBmxo7fb5Md-IY7d_4 state=0x7faf5412a068 s->prefetch_data=0
2013-09-03 18:01:36.590886 7faf46ff5700 20 get_obj_state: s->obj_tag was set empty
2013-09-03 18:01:36.590921 7faf46ff5700 20 prepare_atomic_for_write_impl: state is not atomic. state=0x7faf5412a068
2013-09-03 18:01:36.594977 7faf46ff5700 20 get_obj_state: rctx=0x7faf540251c0 obj=test-han:_shadow__Bw6LkKRYSmAwEh5VBmxo7fb5Md-IY7d_5 state=0x7faf54126468 s->prefetch_data=0
2013-09-03 18:01:36.596140 7faf46ff5700 20 get_obj_state: s->obj_tag was set empty
.
.
hunderts of similar  lines
.
.
.
2013-09-03 18:01:42.210941 7faf46ff5700 20 get_obj_state: rctx=0x7faf540251c0 obj=test-han:_shadow__Bw6LkKRYSmAwEh5VBmxo7fb5Md-IY7d_1038 state=0x7faf54172738 s->prefetch_data=0
2013-09-03 18:01:42.211888 7faf46ff5700 20 get_obj_state: s->obj_tag was set empty
2013-09-03 18:01:42.211947 7faf46ff5700 20 prepare_atomic_for_write_impl: state is not atomic. state=0x7faf54172738
2013-09-03 18:01:42.216903 7faf46ff5700 20 get_obj_state: rctx=0x7faf540251c0 obj=test-han:_shadow__Bw6LkKRYSmAwEh5VBmxo7fb5Md-IY7d_1039 state=0x7faf54172a88 s->prefetch_data=0
2013-09-03 18:01:42.217757 7faf46ff5700 20 prepare_atomic_for_write_impl: state is not atomic. state=0x7faf54172a88
2013-09-03 18:01:42.221839 7faf46ff5700  0 WARNING: set_req_state_err err_no=27 resorting to 500
2013-09-03 18:01:42.221868 7faf46ff5700  2 req 36:402.848669:s3:PUT /test-han/tesfile05.iso:put_obj:http status=500
2013-09-03 18:01:42.223232 7faf46ff5700  1 ====== req done req=0x25e24b0 http_status=500 ======
2013-09-03 18:01:55.427440 7faf87fff700  2 RGWDataChangesLog::ChangesRenewThread: start
2013-09-03 18:02:01.331323 7faf9d011780 20 enqueued request req=0x25e2020
2013-09-03 18:02:01.331341 7faf9d011780 20 RGWWQ:
2013-09-03 18:02:01.331343 7faf9d011780 20 req: 0x25e2020
2013-09-03 18:02:01.331352 7faf9d011780 10 allocated request req=0x25e2500
2013-09-03 18:02:01.331377 7faf307c8700 20 dequeued request req=0x25e2020
2013-09-03 18:02:01.331385 7faf307c8700 20 RGWWQ: empty






3. performance
While I get great performance from the radosgw to the stoage nodes, the performance from cyberduck to the radosgw is poor (10 times slower)

On a rados filesystem mounted on the radosgw I get: arround 100MB/s write troughput (here I know it's the 1Gb sync network between the nodes who is the limiting factor).
With a cyberduck test I only get 10MB/s and I cannot identify the bottleneck, network between testclient and radosgw is 1Gb and verified.


Regards
Andi
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux