Re: Creating a bucket with bucket constructor in Ceph v16.2.7

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Ramin,

I think youre still going to experience what Casey described.

If your intent is to completely isolate bucket metadata/data in one
zonegroup from another, then I believe you need multiple independent
realms. Each with its own endpoint.

For instance;

Ceph Cluster A
Realm1/zonegroup1/zone1 (endpoint.left)
Realm2/zonegroup2/zone2 (endpoint.right)

Then, you dont need to bother with location constraint. Attempts to
create buckets via endpoint.right will be created at realm2/zonegroup2.
Buckets created via endpoint.left will show up at zonegroup1/zone1. They
will be completely isolated from one another, yet reside on the same
cluster.

Location constraint could be used for different placement targets within
zonegroup2. For instance, location constraint zonegroup2:put-data-on-
replicated-storage or zonegroup2:put-data-on-erasure-storage. Without
separate realms, I believe you will continue to experience what Casey
explained.

If it is your intent to have multiple isolated zonegroups within the
same cluster, then I would read through the docs for creating a multi-
realm deployment. And just stop short of creating an additional
synchronized replicated realm. (step #15)

https://www.ibm.com/docs/en/storage-ceph/5?topic=administration-configuring-multiple-realms-in-same-storage-cluster

You should then have two independent realms residing on the same
cluster. You can further isolate the two zonegroups by deploying new
root pools. By default, I believe the zonegroup/zone settings are stored
in .rgw.root

Any actions without specifying an alternative zonegroup/zone will by
default fall under the .rgw.root hierarchy.

You could, for example, deploy two new realms, each with its own root
pool.

zonegroup1 --> .zonegroup1.rgw.root
zonegroup2 --> .zonegroup2.rgw.root 

These two zonegroups will be isolated in their own realms, and also
isolated from the default .rgw.root

Once you have that setup, it's like each realm is its own completely
separate rgw deployment, but reside within a single Ceph cluster. It
takes a bit of forethought to get everything set up correctly. You will
need to create keys for each realm and be mindful of which realm you are
operating on. It may take several attempts to get everything in place
the way you expect it to be.


On Sat, 2023-05-20 at 15:42 +0330, Ramin Najjarbashi wrote:
> I had a typo in s3cmd output :D
> 
> I am encountering an issue with the LocationConstraint when using
> s3cmd to
> create a bucket. In my setup, I have two zone groups within the same
> realm.
> My intention is to create the bucket in the second zone group by
> correctly
> setting the bucket location using bucket_location:
> <CreateBucketConfiguration><LocationConstraint>abrak</LocationConstrai
> nt></CreateBucketConfiguration>.
> However, despite configuring it this way, the bucket is still being
> created
> in the first zone group.
> 
> https://gist.github.com/RaminNietzsche/1ff266a6158b437319dcd2eb10eeb34e
> 
> ```sh
> s3cmd --region zg2-api-name mb s3://test-zg2s3cmd info
> s3://test-zg2s3://test-zg2/ (bucket):
>    Location:  zg1-api-name
>    Payer:     BucketOwner
>    Expiration Rule: none
>    Policy:    none
>    CORS:      none
>    ACL:       development: FULL_CONTROL
> ```
> 
> 
> On Thu, May 18, 2023 at 4:29 PM Ramin Najjarbashi <
> ramin.najarbashi@xxxxxxxxx> wrote:
> 
> > Thank Casey
> > 
> > Currently, when I create a new bucket and specify the bucket
> > location as
> > zone group 2, I expect the request to be handled by the master zone
> > in zone
> > group 1, as it is the expected behavior. However, I noticed that
> > regardless
> > of the specified bucket location, the zone group ID for all buckets
> > created
> > using this method remains the same as zone group 1.
> > 
> > 
> > My expectation was that when I create a bucket in zone group 2, the
> > zone
> > group ID in the bucket metadata would reflect the correct zone group
> > ID.
> > 
> > 
> > 
> > On Thu, May 18, 2023 at 15:54 Casey Bodley <cbodley@xxxxxxxxxx>
> > wrote:
> > 
> > > On Wed, May 17, 2023 at 11:13 PM Ramin Najjarbashi
> > > <ramin.najarbashi@xxxxxxxxx> wrote:
> > > > 
> > > > Hi
> > > > 
> > > > I'm currently using Ceph version 16.2.7 and facing an issue with
> > > > bucket
> > > > creation in a multi-zone configuration. My setup includes two
> > > > zone
> > > groups:
> > > > 
> > > > ZG1 (Master) and ZG2, with one zone in each zone group (zone-1
> > > > in ZG1
> > > and
> > > > zone-2 in ZG2).
> > > > 
> > > > The objective is to create buckets in a specific zone group
> > > > (ZG2) using
> > > the
> > > > bucket constructor.
> > > > However, despite setting the desired zone group (abrak) in the
> > > > request,
> > > the
> > > > bucket is still being created in the master zone group (ZG1).
> > > > I have defined the following endpoint pattern for each zone
> > > > group:
> > > > 
> > > > s3.{zg}.mydomain.com
> > > > 
> > > > I am using the s3cmd client to interact with the Ceph cluster. I
> > > > have
> > > > ensured that I provide the necessary endpoint and region
> > > > information
> > > while
> > > > executing the bucket creation command. Despite my efforts, the
> > > > bucket
> > > > consistently gets created in ZG1 instead of ZG2.
> > > 
> > > this is expected behavior for the metadata consistency model. all
> > > metadata gets created on the metadata master zone first, and syncs
> > > to
> > > all other zones in the realm from there. so your buckets will be
> > > visible to every zonegroup
> > > 
> > > however, ZG2 is still the 'bucket location', and its object data
> > > should only reside in ZG2's zones. any s3 requests on that bucket
> > > sent
> > > to ZG1 will get redirected to ZG2 and serviced there
> > > 
> > > if you don't want any metadata shared between the two zonegroups,
> > > you
> > > can put them in separate realms. but that includes user metadata
> > > as
> > > well
> > > 
> > > > 
> > > > - Ceph Version: 16.2.7
> > > > - Zone Group 1 (ZG1) Endpoint: http://s3.zonegroup1.mydomain.com
> > > > - Zone Group 2 (ZG2) Endpoint: http://s3.zonegroup2.mydomain.com
> > > > - Desired Bucket Creation Region: zg2-api-name
> > > > 
> > > >  have reviewed the Ceph documentation and made necessary
> > > > configuration
> > > > changes, but I have not been able to achieve the desired
> > > > result.I kindly
> > > > request your assistance in understanding why the bucket
> > > > constructor is
> > > not
> > > > honoring the specified region and always defaults to ZG1. I
> > > > would
> > > greatly
> > > > appreciate any insights, recommendations, or potential solutions
> > > > to
> > > resolve
> > > > this issue.
> > > > 
> > > >  Thank you for your time and support.
> > > > 
> > > > ---------------------------------
> > > > Here are the details of my setup:
> > > > ---------------------------------
> > > > 
> > > > ```sh
> > > > s3cmd --region zg2-api-name mb s3://test-zg2s3cmd info
> > > > s3://test-zg2s3://test-zg2/ (bucket):
> > > >    Location:  zg2-api-name
> > > >    Payer:     BucketOwner
> > > >    Expiration Rule: none
> > > >    Policy:    none
> > > >    CORS:      none
> > > >    ACL:       development: FULL_CONTROL
> > > > ```
> > > > 
> > > > this is my config file:
> > > > 
> > > > ```ini
> > > > [default]
> > > > access_key = <KEY>
> > > > secret_key = <SECRET>
> > > > host_base = s3.zonegroup1.mydomain.com
> > > > host_bucket = s3.%(location)s.mydomain.com
> > > > #host_bucket = %(bucket)s.s3.zonegroup1.mydomain.com
> > > > #host_bucket = s3.%(location)s.mydomain.com
> > > > #host_bucket = s3.%(region)s.mydomain.com
> > > > bucket_location = zg1-api-name
> > > > use_https = False
> > > > ```
> > > > 
> > > > 
> > > > Zonegroup configuration for the `zonegroup1` region:
> > > > 
> > > > ```json
> > > > {
> > > >     "id": "fb3f818a-ca9b-4b12-b431-7cdcd80006d",
> > > >     "name": "zg1-api-name",
> > > >     "api_name": "zg1-api-name",
> > > >     "is_master": "false",
> > > >     "endpoints": [
> > > >         "http://s3.zonegroup1.mydomain.com";,
> > > >     ],
> > > >     "hostnames": [
> > > >         "s3.zonegroup1.mydomain.com",
> > > >     ],
> > > >     "hostnames_s3website": [
> > > >         "s3-website.zonegroup1.mydomain.com",
> > > >     ],
> > > >     "master_zone": "at2-stg-zone",
> > > >     "zones": [
> > > >         {
> > > >             "id": "at2-stg-zone",
> > > >             "name": "at2-stg-zone",
> > > >             "endpoints": [
> > > >                 "http://s3.zonegroup1.mydomain.com";
> > > >             ],
> > > >             "log_meta": "false",
> > > >             "log_data": "true",
> > > >             "bucket_index_max_shards": 11,
> > > >             "read_only": "false",
> > > >             "tier_type": "",
> > > >             "sync_from_all": "true",
> > > >             "sync_from": [],
> > > >             "redirect_zone": ""
> > > >         }
> > > >     ],
> > > >     "placement_targets": [
> > > >         {
> > > >             "name": "default-placement",
> > > >             "tags": [],
> > > >             "storage_classes": [
> > > >                 "STANDARD"
> > > >             ]
> > > >         }
> > > >     ],
> > > >     "default_placement": "default-placement",
> > > >     "realm_id": "fa2f8194-4a9d-4b98-b411-9cdcd1e5506a",
> > > >     "sync_policy": {
> > > >         "groups": []
> > > >     }
> > > > }
> > > > ```
> > > > 
> > > > Zonegroup configuration for the `zonegroup2` region:
> > > > 
> > > > ```json
> > > > {
> > > >     "id": "a513d60c-44a2-4289-a23d-b7a511be6ee4",
> > > >     "name": "zg2-api-name",
> > > >     "api_name": "zg2-api-name",
> > > >     "is_master": "false",
> > > >     "endpoints": [
> > > >         "http://s3.zonegroup2.mydomain.com";
> > > >     ],
> > > >     "hostnames": [
> > > >         "s3.zonegroup2.mydomain.com"
> > > >     ],
> > > >     "hostnames_s3website": [],
> > > >     "master_zone": "zonegroup2-sh-1",
> > > >     "zones": [
> > > >         {
> > > >             "id": "zonegroup2-sh-1",
> > > >             "name": "zonegroup2-sh-1",
> > > >             "endpoints": [
> > > >                 "http://s3.zonegroup2.mydomain.com";
> > > >             ],
> > > >             "log_meta": "false",
> > > >             "log_data": "false",
> > > >             "bucket_index_max_shards": 11,
> > > >             "read_only": "false",
> > > >             "tier_type": "",
> > > >             "sync_from_all": "true",
> > > >             "sync_from": [],
> > > >             "redirect_zone": ""
> > > >         }
> > > >     ],
> > > >     "placement_targets": [
> > > >         {
> > > >             "name": "default-placement",
> > > >             "tags": [],
> > > >             "storage_classes": [
> > > >                 "STANDARD"
> > > >             ]
> > > >         }
> > > >     ],
> > > >     "default_placement": "default-placement",
> > > >     "realm_id": "fa2f8194-4a9d-4b98-b411-9cdcd1e5506a",
> > > >     "sync_policy": {
> > > >         "groups": []
> > > >     }
> > > > }
> > > > ```
> > > > _______________________________________________
> > > > ceph-users mailing list -- ceph-users@xxxxxxx
> > > > To unsubscribe send an email to ceph-users-leave@xxxxxxx
> > > > 
> > > 
> > > 
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
> 

Attachment: smime.p7s
Description: S/MIME cryptographic signature

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux