Re: radosgw-agent testing

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



> On Thu, Aug 22, 2013 at 4:38 AM, christophe courtaut
> <christophe.courtaut@xxxxxxxxx> wrote:
>> Hi Yehuda,
>>
>> I'm currently trying to test the radosgw-agent which lives in
>> https://github.com/ceph/radosgw-agent
>>
>> I modified the vstart.sh script (you can see it here
>> https://github.com/ceph/ceph/pull/522) to be able to launch multiple
>> cluster with vstart.sh in different directories.
>>
>> I created a little script to automate the launch of two clusters and
>> the setup of region and zone, so that i have a master and a slave
>> cluster.
>>
>> With that setup, i encounter two problems:
>>
>> First when i try to create a bucket after setup is done using s3cmd,
>> got this error
>> ERROR: Access to bucket 'TEST' was denied.
>> in the function open_bucket_pool_ctx, it tries to open the pool
>> .rgw.buckets.index, but this pool doesn't seems to exist.
>> Is this behaviour correct? Is it due to my setup? or should this pool
>> be created?
>
> This pool should have been automatically created (if didn't exist
> before). Not sure why it's not happening.

Ok. I don't know where the automatic creation of this pool should be
handled though.

>>
>> Second, is that i get a 403 while using the radosgw-agent on the
>> previous setup, with the test.sh script.
>> Am i doing something wrong here?
>
> Did you set the 'system' flag on the appropriate users?

No i did miss that point. btw the option was not mentionned in the
radosgw-admin help.
here comes a pull request for that : https://github.com/ceph/ceph/pull/528

>> To reproduce the first bug:
>>
>> - Get the vstart.sh script from https://github.com/ceph/ceph/pull/522
>> - Get the script to launch cluster and the *.master *.zone files
>> - Launch the script
>> - Try to create a bucket with s3cmd
>>
>> To reproduce the second one :
>>
>> - Do the same setup as for the first one
>> - Get the radosgw-agent repo
>> - Get the test.sh script in this directory
>> - Launch the test.sh script
>>
>> You can find below the various files i mentionned earlier
>>
>> the script for launching clusters
>> http://pastebin.com/VSJhwRSs
>>
>> master.region file
>> http://pastebin.com/pCrRpM3w
>>
>> master.zone file
>> http://pastebin.com/gi1DDNvr
>>
>> slave.region file
>> http://pastebin.com/idgMDYjj
>>
>> slave.zone file
>> http://pastebin.com/m2KRnqdJ
>
> On both slave and master zone you didn't set the placement pools, so
> it resorts to the default. What does 'radosgw-admin zone get' show on
> both?

radosgw-admin -c cluster-master/ceph.conf zone get

{ "domain_root": ".rgw",
  "control_pool": ".rgw.control",
  "gc_pool": ".rgw.gc",
  "log_pool": ".log",
  "intent_log_pool": ".intent-log",
  "usage_log_pool": ".usage",
  "user_keys_pool": ".users",
  "user_email_pool": ".users.email",
  "user_swift_pool": ".users.swift",
  "user_uid_pool": ".",
  "system_key": { "access_key": "0555b35654ad1656d804",
      "secret_key":
"h7GhxuBLTrlhVUyxSPUKUV8r\/2EI4ngqJxD7iBdBYLhwluN30JaT3Q=="},
  "placement_pools": []}

radosgw-admin -c cluster-slave/ceph.conf zone get

{ "domain_root": ".rgw",
  "control_pool": ".rgw.control",
  "gc_pool": ".rgw.gc",
  "log_pool": ".log",
  "intent_log_pool": ".intent-log",
  "usage_log_pool": ".usage",
  "user_keys_pool": ".users",
  "user_email_pool": ".users.email",
  "user_swift_pool": ".users.swift",
  "user_uid_pool": ".",
  "system_key": { "access_key": "0555b35654ad1656d804",
      "secret_key":
"h7GhxuBLTrlhVUyxSPUKUV8r\/2EI4ngqJxD7iBdBYLhwluN30JaT3Q=="},
  "placement_pools": []}

Indeed placement_pools seems empty.
Is this mandatory?
If so do i need to create the pools first?
On Both cluster?
Should they be the same?

>>
>> test.sh file
>> http://pastebin.com/gaitrPqX
>
> You can put most of these configurables in a yaml file and -c conf.yaml.

Ok thanks. i'll do that!

-- 
Christophe Courtaut
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux