Hi All, I saw the previous thread on this related to http://tracker.ceph.com/issues/15597 and Yehuda's fix script https://raw.githubusercontent.com/yehudasa/ceph/wip-fix-default-zone/src/fix-zone Running this seems to have landed me in a weird state. I can create and get new buckets and objects but I've "lost" all my old buckets. I'm fairly confident the "lost" data is in the .rgw.buckets pool but my current zone is set to use .rgw.buckets_ root@ceph-mon0:~# radosgw-admin zone get { "id": "default", "name": "default", "domain_root": ".rgw_", "control_pool": ".rgw.control_", "gc_pool": ".rgw.gc_", "log_pool": ".log_", "intent_log_pool": ".intent-log_", "usage_log_pool": ".usage_", "user_keys_pool": ".users_", "user_email_pool": ".users.email_", "user_swift_pool": ".users.swift_", "user_uid_pool": ".users.uid_", "system_key": { "access_key": "", "secret_key": "" }, "placement_pools": [ { "key": "default-placement", "val": { "index_pool": ".rgw.buckets.index_", "data_pool": ".rgw.buckets_", "data_extra_pool": ".rgw.buckets.extra_", "index_type": 0 } } ], "metadata_heap": "default.rgw.meta", "realm_id": "a935d12f-14b7-4bf8-a24f-596d5ddd81be" } root@ceph-mon0:~# ceph osd pool ls |grep rgw|sort default.rgw.meta .rgw .rgw_ .rgw.buckets .rgw.buckets_ .rgw.buckets.index .rgw.buckets.index_ .rgw.control .rgw.control_ .rgw.gc .rgw.gc_ .rgw.root .rgw.root.backup Should I just adjust the zone to use the pools without trailing slashes? I'm a bit lost. the last I could see from running the script didn't seem to indicate any errors (though I lost the to to scroll back buffer before i noticed the issue) Tail of output from running script: https://raw.githubusercontent.com/yehudasa/ceph/wip-fix-default-zone/src/fix-zone + radosgw-admin zone set --rgw-zone=default zone id default{ "id": "default", "name": "default", "domain_root": ".rgw_", "control_pool": ".rgw.control_", "gc_pool": ".rgw.gc_", "log_pool": ".log_", "intent_log_pool": ".intent-log_", "usage_log_pool": ".usage_", "user_keys_pool": ".users_", "user_email_pool": ".users.email_", "user_swift_pool": ".users.swift_", "user_uid_pool": ".users.uid_", "system_key": { "access_key": "", "secret_key": "" }, "placement_pools": [ { "key": "default-placement", "val": { "index_pool": ".rgw.buckets.index_", "data_pool": ".rgw.buckets_", "data_extra_pool": ".rgw.buckets.extra_", "index_type": 0 } } ], "metadata_heap": "default.rgw.meta", "realm_id": "a935d12f-14b7-4bf8-a24f-596d5ddd81be" } + radosgw-admin zonegroup default --rgw-zonegroup=default + radosgw-admin zone default --rgw-zone=default root@ceph-mon0:~# radosgw-admin region get --rgw-zonegroup=default { "id": "default", "name": "default", "api_name": "", "is_master": "true", "endpoints": [], "hostnames": [], "hostnames_s3website": [], "master_zone": "default", "zones": [ { "id": "default", "name": "default", "endpoints": [], "log_meta": "false", "log_data": "false", "bucket_index_max_shards": 0, "read_only": "false" } ], "placement_targets": [ { "name": "default-placement", "tags": [] } ], "default_placement": "default-placement", "realm_id": "a935d12f-14b7-4bf8-a24f-596d5ddd81be"} root@ceph-mon0:~# ceph -v ceph version 10.2.1 (3a66dd4f30852819c1bdaa8ec23c795d4ad77269) Thanks, -Jon _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com