I'll try the pro version of crossftp as soon as I have a chance. Here is the output using s3cmd version 1.1.0-beta3: root@theneykov:/mnt/samba-rbd/Vantage/Incoming/ascvid# s3cmd -v put --debug 2 20130718_ascvid_cheyennemizeTEST4.mov s3://linux DEBUG: ConfigParser: Reading file '/root/.s3cfg' DEBUG: ConfigParser: access_key->PK...17_chars...F DEBUG: ConfigParser: bucket_location->US DEBUG: ConfigParser: cloudfront_host->cloudfront.amazonaws.com DEBUG: ConfigParser: default_mime_type->binary/octet-stream DEBUG: ConfigParser: delete_removed->False DEBUG: ConfigParser: dry_run->False DEBUG: ConfigParser: enable_multipart->True DEBUG: ConfigParser: encoding->UTF-8 DEBUG: ConfigParser: encrypt->False DEBUG: ConfigParser: follow_symlinks->False DEBUG: ConfigParser: force->False DEBUG: ConfigParser: get_continue->False DEBUG: ConfigParser: gpg_command->/usr/bin/gpg DEBUG: ConfigParser: gpg_decrypt->%(gpg_command)s -d --verbose --no-use-agent --batch --yes --passphrase-fd %(passphrase_fd)s -o %(output_file)s %(input_file)s DEBUG: ConfigParser: gpg_encrypt->%(gpg_command)s -c --verbose --no-use-agent --batch --yes --passphrase-fd %(passphrase_fd)s -o %(output_file)s %(input_file)s DEBUG: ConfigParser: gpg_passphrase->...-3_chars... DEBUG: ConfigParser: guess_mime_type->True DEBUG: ConfigParser: host_base->nprs3.com DEBUG: ConfigParser: host_bucket->%(bucket)s.nprs3.com DEBUG: ConfigParser: human_readable_sizes->False DEBUG: ConfigParser: invalidate_on_cf->False DEBUG: ConfigParser: list_md5->False DEBUG: ConfigParser: log_target_prefix-> DEBUG: ConfigParser: mime_type-> DEBUG: ConfigParser: multipart_chunk_size_mb->15 DEBUG: ConfigParser: preserve_attrs->True DEBUG: ConfigParser: progress_meter->True DEBUG: ConfigParser: proxy_host-> DEBUG: ConfigParser: proxy_port->0 DEBUG: ConfigParser: recursive->False DEBUG: ConfigParser: recv_chunk->4096 DEBUG: ConfigParser: reduced_redundancy->False DEBUG: ConfigParser: secret_key->J7...37_chars...Y DEBUG: ConfigParser: send_chunk->4096 DEBUG: ConfigParser: simpledb_host->sdb.amazonaws.com DEBUG: ConfigParser: skip_existing->False DEBUG: ConfigParser: socket_timeout->300 DEBUG: ConfigParser: urlencoding_mode->normal DEBUG: ConfigParser: use_https->False DEBUG: ConfigParser: verbosity->WARNING DEBUG: ConfigParser: website_endpoint->http://%(bucket)s.s3-website-%(location)s.amazonaws.com/ DEBUG: ConfigParser: website_error-> DEBUG: ConfigParser: website_index->index.html DEBUG: Updating Config.Config encoding -> UTF-8 DEBUG: Updating Config.Config follow_symlinks -> False DEBUG: Updating Config.Config verbosity -> 10 DEBUG: Unicodising 'put' using UTF-8 DEBUG: Unicodising '2' using UTF-8 DEBUG: Unicodising '20130718_ascvid_cheyennemizeTEST4.mov' using UTF-8 DEBUG: Unicodising 's3://linux' using UTF-8 DEBUG: Command: put INFO: Compiling list of local files... DEBUG: DeUnicodising u'' using UTF-8 DEBUG: DeUnicodising u'2' using UTF-8 INFO: Compiling list of local files... DEBUG: DeUnicodising u'' using UTF-8 DEBUG: DeUnicodising u'20130718_ascvid_cheyennemizeTEST4.mov' using UTF-8 DEBUG: Unicodising '20130718_ascvid_cheyennemizeTEST4.mov' using UTF-8 DEBUG: Unicodising '20130718_ascvid_cheyennemizeTEST4.mov' using UTF-8 INFO: Applying --exclude/--include DEBUG: CHECK: 20130718_ascvid_cheyennemizeTEST4.mov DEBUG: PASS: 20130718_ascvid_cheyennemizeTEST4.mov INFO: Summary: 1 local files to upload DEBUG: Unicodising '20130718_ascvid_cheyennemizeTEST4.mov' using UTF-8 WARNING: File can not be uploaded: 20130718_ascvid_cheyennemizeTEST4.mov: Permission denied We will need a command line tool to use in cronjobs, etc...so I am hoping I can this working soon. Thanks agin for the help already. Shain Shain Miley | Manager of Systems and Infrastructure, Digital Media | smiley@xxxxxxx | 202.513.3649 ________________________________________ From: Derek Yarnell [derek@xxxxxxxxxxxxxx] Sent: Saturday, October 26, 2013 6:46 PM To: Shain Miley; ceph-users@xxxxxxxx Subject: Re: Radosgw and large files Hi Shain, Yes we have tested and have working S3 Multipart support for files >5GB (RHEL64/0.67.4). However, crossftp unless you have pro it would seem does not support multipart. Dragondisk gives the error that I have seen when using a PUT and not multipart, EntityTooLarge. My guess is that it is not doing Multipart. Now from s3cmd's source it does support multipart have you tried to give it a --debug 2 which will output boto's debug output I believe. That should tell you more about if it is correctly doing a multipart upload. Thanks, derek On 10/26/13, 5:10 PM, Shain Miley wrote: > Hi, > > I am wondering if anyone has successfully been able to upload files > larger than 5GB using radosgw. > > I have tried using various clients, including dragondisk, crossftp, > s3cmd, etc...all of them have failed with a 'permission denied' response. > > Each of the clients say they support multi-part (and it appears that > they do as files larger than 15MB gets split into multiple pieces) > however I can only upload files smaller than 5GB up to this point. > > I am using 0.67.4 and Ubuntu 12.04. > > Thanks in advance, > > Shain > > Shain Miley | Manager of Systems and Infrastructure, Digital Media | > smiley@xxxxxxx | 202.513.3649 > > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > -- --- Derek T. Yarnell University of Maryland Institute for Advanced Computer Studies _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com