I haven't ran rgw over nginx for quite a while, so I'm not sure whether it can actually work. The problem that you're seeing might be related to the 100-continue processing, which is now enabled by default. Try to turn it off, setting the following under the global (or client) section in your ceph.conf: rgw print continue = false If it doesn't help we'll need to dig deeper. Thanks, Yehuda 2011/11/8 Sławomir Skowron <szibis@xxxxxxxxx>: > Maybe, i have forgot something, but there is no doc about that. > > I create a configuration with nginx and radosgw for S3. > > On top of radosgw standing nginx witch cache capability. Everything > was ok in version 0.32 of ceph. A have create a new filesystem with a > newest 0.37 version, and now i have some problems. > > I run radosgw like this: > > radosgw --rgw-socket-path=/var/run/radosgw.sock --conf=/etc/ceph/ceph.conf > > In nginx i talk to unix socket of radosgw. Everything looks good. In > Radosgw-admin i create user, and its ok. > > { "user_id": "0", > "rados_uid": 0, > "display_name": "ocd", > "email": "", > "suspended": 0, > "subusers": [], > "keys": [ > { "user": "0", > "access_key": "CFLZFEPYUAZV4EZ1P8OJ", > "secret_key": "HrWN8SNfjXPhUELPLIbRIA3nCfppQjJ5xV6EnhNM"}], > > pool name category KB objects clones > degraded unfound rd rd KB wr > wr KB > .intent-log - 1 1 0 > 0 0 0 0 1 > 1 > .log - 19 18 0 > 0 0 0 0 39 > 39 > .rgw - 20 25 0 > 0 0 1 0 49 > 46 > .rgw.buckets - 19 19 0 > 0 0 32 15 45 > 23 > .users - 2 2 0 > 0 0 0 0 2 > 2 > .users.email - 1 1 0 > 0 0 0 0 1 > 1 > .users.uid - 4 4 0 > 0 0 3 1 5 > 5 > data - 0 0 0 > 0 0 0 0 0 > 0 > metadata - 0 0 0 > 0 0 0 0 0 > 0 > rbd - 0 0 0 > 0 0 0 0 0 > 0 > total used 540000 71 > total avail 174920736 > total space 175460736 > > But when i testing this with a local s3lib its not working. In nginx acces log: > > 127.0.0.1 - - - [08/Nov/2011:09:03:48 +0100] "PUT /nodejs/test01 > HTTP/1.1" rlength: 377 bsent: 270 rtime: 0.006 urtime: 0.004 status: > 403 bbsent: 103 httpref: "-" useragent: "Mozilla/4.0 (Compatible; s3; > libs3 2.0; Linux x86_64)" > 127.0.0.1 - - - [08/Nov/2011:09:03:55 +0100] "PUT /nodejs/test01 > HTTP/1.1" rlength: 377 bsent: 270 rtime: 0.006 urtime: 0.004 status: > 403 bbsent: 103 httpref: "-" useragent: "Mozilla/4.0 (Compatible; s3; > libs3 2.0; Linux x86_64)" > > Request have code 100, waiting for something in radosgw. > > s3 -u put nodejs/test01 < /usr/src/libs3-2.0/TODO > > ERROR: ErrorAccessDenied > > From external s3 client, a have something like that. Somehow bucket > was created, but other operations not working with error access > denied. > > <?xml version='1.0' encoding='UTF-8'?> > <Error> > <Code>AccessDenied</Code> > </Error> > > I have some data from s3lib, by a many tries: > > Bucket Created > -------------------------------------------------------- -------------------- > nodejs 2011-11-07T13:11:42Z > root@vm-10-177-48-24:/usr/src# s3 -u getacl nodejs > OwnerID 0 ocd > Type User Identifier > Permission > ------ ------------------------------------------------------------------------------------------ > ------------ > UserID 0 (ocd) > FULL_CONTROL > root@vm-10-177-48-24:/usr/src# s3 -u test nodejs > Bucket Status > -------------------------------------------------------- -------------------- > nodejs USA > root@vm-10-177-48-24:/usr/src# s3 -u getacl nodejs > OwnerID 0 ocd > Type User Identifier > Permission > ------ ------------------------------------------------------------------------------------------ > ------------ > UserID 0 (ocd) > FULL_CONTROL > root@vm-10-177-48-24:/usr/src# s3 -u list > Bucket Created > -------------------------------------------------------- -------------------- > nodejs 2011-11-07T13:11:42Z > > > Ceph.conf > > ; global > [global] > ; enable secure authentication > auth supported = cephx > keyring = /etc/ceph/keyring.bin > > ; monitors > ; You need at least one. You need at least three if you want to > ; tolerate any node failures. Always create an odd number. > [mon] > mon data = /vol0/data/mon.$id > > ; some minimal logging (just message traffic) to aid debugging > > debug ms = 1 ; see message traffic > debug mon = 0 ; monitor > debug paxos = 0 ; monitor replication > debug auth = 0 ; > > mon allowed clock drift = 2 > > [mon.0] > host = vm-10-177-48-24 > mon addr = 10.177.48.24:6789 > > ; osd > ; You need at least one. Two if you want data to be replicated. > ; Define as many as you like. > [osd] > ; This is where the btrfs volume will be mounted. > osd data = /vol0/data/osd.$id > > ; Ideally, make this a separate disk or partition. A few GB > ; is usually enough; more if you have fast disks. You can use > ; a file under the osd data dir if need be > ; (e.g. /data/osd$id/journal), but it will be slower than a > ; separate disk or partition. > osd journal = /vol0/data/osd.$id/journal > > ; If the OSD journal is a file, you need to specify the size. > This is specified in MB. > osd journal size = 512 > > filestore journal writeahead = 1 > osd heartbeat grace = 5 > > debug ms = 1 ; message traffic > debug osd = 0 > debug filestore = 0 ; local object storage > debug journal = 0 ; local journaling > debug monc = 0 > debug rados = 0 > > [osd.0] > host = vm-10-177-48-24 > osd data = /vol0/data/osd.0 > keyring = /vol0/data/osd.0/keyring > > [osd.1] > host = vm-10-177-48-24 > osd data = /vol0/data/osd.1 > keyring = /vol0/data/osd.1/keyring > > > radosgw-admin bucket stats --bucket=nodejs > { "bucket": "nodejs", > "pool": ".rgw.buckets", > "id": 10, > "marker": "10", > "owner": "0", > "usage": { "rgw.main": { "size_kb": 4, > "num_objects": 1}, > "rgw.shadow": { "size_kb": 4, > "num_objects": 1}}} > > ceph -s > 2011-11-08 09:07:12.546715 pg v594: 460 pgs: 460 active+clean; 68 > KB data, 527 MB used, 166 GB / 167 GB avail > 2011-11-08 09:07:12.547555 mds e1: 0/0/1 up > 2011-11-08 09:07:12.547573 osd e12: 2 osds: 2 up, 2 in > 2011-11-08 09:07:12.547626 log 2011-11-08 08:31:46.761320 osd.0 > 10.177.48.24:6800/12063 244 : [INF] 10.3 scrub ok > 2011-11-08 09:07:12.547709 mon e1: 1 mons at {0=10.177.48.24:6789/0} > > > -- > ----- > Pozdrawiam > > Sławek "sZiBis" Skowron > -- > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html > ��.n��������+%������w��{.n����z��u���ܨ}���Ơz�j:+v�����w����ޙ��&�)ߡ�a����z�ޗ���ݢj��w�f