Hi all, I am trying to add a third zone to our zonegroup with the tier type archive (the archive sync module). However this caused the archive Ceph cluster to fill up entirely and made radosgw-admin unusable. I had to delete the zone pools and delete the zone to get the cluster up and running again without adding more disks. There are a lot of errors on the RGW server logs like: RGW-SYNC:data:sync:shard[37]: a sync operation returned error data sync: SYNC_ARCHIVE: sync_object: error versioning archive bucket And the following error messages from *radosgw-admin sync error list*: failed to sync bucket instance: (125) Operation canceled failed to read remote metadata entry: (5) Input/output error I have now updated the Ceph cluster from 14.2.16 to 14.2.22 and recreated the zone. The sync is currently running again but still with the same errors. The output of *radosgw-admin sync status* is: radosgw-admin sync status realm 363d82c9-72df-4c82-b8fa-4196e3307fe5 (company) zonegroup f691f801-992f-49d0-8a7d-39d1b73eeea8 (fo) zone cbe6a428-591e-4a94-96ee-dcb18fc30eeb (fo-of) metadata sync syncing full sync: 0/64 shards incremental sync: 64/64 shards metadata is caught up with master data sync source: dffdf44a-6d0d-4ded-bd29-f716d31b1a2d (fo-am) syncing full sync: 109/128 shards full sync: 7809 buckets to sync incremental sync: 19/128 shards data is behind on 120 shards behind shards: [0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,45,46,47,48,49,50,52,53,54,55,56,57,58,60,61,62,63,64,65,66,67,68,69,70,71,72,73,75,76,77,78,79,80,81,82,83,84,85,86,87,88,89,91,92,93,94,95,96,97,98,99,100,101,102,103,104,105,107,108,109,110,111,112,113,114,115,116,117,118,119,120,121,123,125,126,127] oldest incremental change not applied: 2021-07-26 11:08:25.0.984823s [115] 14 shards are recovering recovering shards: [44,51,58,59,60,74,75,76,90,91,92,106,122,124] source: f4b0e442-1b70-4bdd-b019-9827f181a6af (fo-ch) syncing full sync: 103/128 shards full sync: 7080 buckets to sync incremental sync: 25/128 shards data is behind on 114 shards behind shards: [0,1,2,4,5,6,7,8,9,11,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58,59,60,61,62,63,64,65,66,67,68,69,70,71,72,73,77,78,79,80,81,82,84,85,86,87,88,89,91,93,94,95,96,97,98,100,101,102,103,104,105,106,109,110,111,112,113,114,115,116,117,118,119,120,121,122,124,125,126,127] oldest incremental change not applied: 2021-07-26 11:08:11.0.648844s [115] 12 shards are recovering recovering shards: [10,11,12,35,51,67,83,92,99,106,107,108] And that makes me wonder if the archive tries to sync the same data from both zones? Don't I need to configure sync_from_all=false the sync_from to one of the other zones? I can't find anything about this in the documentation or by searching online. Kind regards, Yosh de Vos _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx