On 03/03/2017 07:40 AM, K K wrote: > > Hello, all! > > I have successfully create 2 zone cluster(se and se2). But my radosgw > machines are sending many GET /admin/log requests to each other after > put 10k items to cluster via radosgw. It's look like: > > 2017-03-03 17:31:17.897872 7f21b9083700 1 civetweb: 0x7f222001f660: > 10.30.18.24 - - [03/Mar/2017:17:31:17 +0500] "GET /admin/log/ > HTTP/1.1" 200 0 - - > 2017-03-03 17:31:17.944212 7f21ca0a5700 1 civetweb: 0x7f2200015510: > 10.30.18.24 - - [03/Mar/2017:17:31:17 +0500] "GET /admin/log/ > HTTP/1.1" 200 0 - - > 2017-03-03 17:31:17.945363 7f21b9083700 1 civetweb: 0x7f222001f660: > 10.30.18.24 - - [03/Mar/2017:17:31:17 +0500] "GET /admin/log/ > HTTP/1.1" 200 0 - - > 2017-03-03 17:31:17.988330 7f21ca0a5700 1 civetweb: 0x7f2200015510: > 10.30.18.24 - - [03/Mar/2017:17:31:17 +0500] "GET /admin/log/ > HTTP/1.1" 200 0 - - > 2017-03-03 17:31:18.005993 7f21b9083700 1 civetweb: 0x7f222001f660: > 10.30.18.24 - - [03/Mar/2017:17:31:17 +0500] "GET /admin/log/ > HTTP/1.1" 200 0 - - > 2017-03-03 17:31:18.006234 7f21c689e700 1 civetweb: 0x7f221c011260: > 10.30.18.24 - - [03/Mar/2017:17:31:17 +0500] "GET /admin/log/ > HTTP/1.1" 200 0 - - > > up to 2k rps!!! Do anybody know what is it??? > These are the radosgw instances polling each other for changes to their metadata and data logs. Each log has many shards (64 or 128), and we poll on each shard separately, which generates a lot of small requests. I would expect these requests to be more frequent during heavy load, and less frequent when idle. We don't currently do any throttling on these connections, though we really should. Casey > Tcpdump show the request is: > > GET > /admin/log/?type=data&id=100&info&rgwx-zonegroup=bfe2e3bb-2040-4b1a-9ccb-ab5347ce3017 > HTTP/1.1 > Host: se2.local > Accept: */* > Transfer-Encoding: chunked > AUTHORIZATION: AWS hEY2W7nW3tdodGrsnrdv:v6+m2FGGhqCSDQteGJ4w039X1uw= > DATE: Fri Mar 3 12:32:20 2017 > Expect: 100-continue > > and answer: > > ...2...m{"marker":"1_1488542463.536646_1448.1","last_update":"2017-03-03 > 12:01:03.536646Z"} > > > > All system install on: > OS: Ubuntu 16.04 > ceph version 10.2.5 (c461ee19ecbc0c5c330aca20f7392c9a00730367) > > rga sync status > 2017-03-03 17:36:20.146017 7f7a72b5ea00 0 error in read_id for id : > (2) No such file or directory > 2017-03-03 17:36:20.147015 7f7a72b5ea00 0 error in read_id for id : > (2) No such file or directory > realm d9ed5678-5734-4609-bf7a-fe3d5f700b23 (s) > zonegroup bfe2e3bb-2040-4b1a-9ccb-ab5347ce3017 (se) > zone 9b212551-a7cf-4aaa-9ef6-b18a31a6e032 (se-k8) > metadata sync no sync (zone is master) > data sync source: 029e0f49-f4dc-4f29-8855-bcc23a8bbcd9 (se2-k12) > syncing > full sync: 0/128 shards > incremental sync: 128/128 shards > data is caught up with source > > > > My config files are: > > [client.radosgw.se2-k12-2] > rgw data = /var/lib/ceph/radosgw/ceph-radosgw.se2-k12-2 > rgw zonegroup = se > rgw zone = se2-k12 > #rgw zonegroup root pool = se.root > #rgw zone root pool = se.root > keyring = /etc/ceph/bak.client.radosgw.se2-k12-2.keyring > rgw host = cbrgw04 > rgw dns name = se2.local > log file = /var/log/radosgw/client.radosgw.se2-k12-2.log > rgw_frontends = "civetweb num_threads=50 port=80" > rgw cache lru size = 10 > rgw cache enabled = false > #debug rgw = 20 > rgw enable ops log = false > #log to stderr = false > rgw enable usage log = false > rgw swift versioning enabled = true > rgw swift url = http://se2.local/ > rgw override bucket index max shards = 20 > rgw print continue = false > > > [client.radosgw.se-k8-2] > rgw data = /var/lib/ceph/radosgw/ceph-radosgw.se-k8-2 > rgw zonegroup = se > rgw zone = se-k8 > #rgw zonegroup root pool = .se.root > #rgw zone root pool = .se.root > keyring = /etc/ceph/ceph.client.radosgw.se-k8-2.keyring > rgw host = cnrgw02 > rgw dns name = se.local > log file = /var/log/radosgw/client.radosgw.se-k8-2.log > rgw_frontends = "civetweb num_threads=100 port=80" > rgw cache enabled = false > rgw cache lru size = 10 > #debug rgw = 20 > rgw enable ops log = false > #log to stderr = false > rgw enable usage log = false > rgw swift versioning enabled = true > rgw swift url = http://se.local > rgw override bucket index max shards = 20 > rgw print continue = false > > rga zonegroup get > { > "id": "bfe2e3bb-2040-4b1a-9ccb-ab5347ce3017", > "name": "se", > "api_name": "se", > "is_master": "true", > "endpoints": [ > "http:\/\/se.local:80" > ], > "hostnames": [], > "hostnames_s3website": [], > "master_zone": "9b212551-a7cf-4aaa-9ef6-b18a31a6e032", > "zones": [ > { > "id": "029e0f49-f4dc-4f29-8855-bcc23a8bbcd9", > "name": "se2-k12", > "endpoints": [ > "http:\/\/se2.local:80" > ], > "log_meta": "false", > "log_data": "true", > "bucket_index_max_shards": 0, > "read_only": "false" > }, > { > "id": "9b212551-a7cf-4aaa-9ef6-b18a31a6e032", > "name": "se-k8", > "endpoints": [ > "http:\/\/se.local:80" > ], > "log_meta": "true", > "log_data": "true", > "bucket_index_max_shards": 0, > "read_only": "false" > } > ], > "placement_targets": [ > { > "name": "default-placement", > "tags": [] > } > ], > "default_placement": "default-placement", > "realm_id": "d9ed5678-5734-4609-bf7a-fe3d5f700b23" > } > > > > -- > K K > > > > _______________________________________________ > ceph-users mailing list > ceph-users at lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20170306/b00399b1/attachment.htm>