Hi,
We're experiencing some issues with our radosgw setup. Today we tried to copy bunch of objects between two separate clusters (using our own tool built on top of java s3 api).
All went smooth until we start copying large objects (200G+). We can see that our code handles this case correctly and started multipart upload (s3.initiateMultipartUpload), then it uploaded all the parts in serial mode (s3.uploadPart) and finally completed upload (s3.completeMultipartUpload).
When we've checked consistency of two clusters we found that we have a lot of zero-sized objects (which turns to be our large objects).
I've made more verbose log from radosgw:
two requests (put_obj, complete_multipart) - https://gist.github.com/anonymous/840e0aee5a7ce0326368 (all finished with 200)
radosgw-admin object stat output: https://gist.github.com/anonymous/2b6771bbbad3021364e2
We've tried to upload these objects several times without any luck.
# radosgw --version
ceph version 0.87 (c51c8f9d80fa4e0168aa52685b8de40e42758578)
Thanks in advance.
--
Best regards,
Gleb M Borisov
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com