Re: Radosgw and large files

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I'll try the pro version of crossftp as soon as I have a chance.

Here is the output using s3cmd version 1.1.0-beta3:

root@theneykov:/mnt/samba-rbd/Vantage/Incoming/ascvid# s3cmd -v put --debug 2 20130718_ascvid_cheyennemizeTEST4.mov s3://linux
DEBUG: ConfigParser: Reading file '/root/.s3cfg'
DEBUG: ConfigParser: access_key->PK...17_chars...F
DEBUG: ConfigParser: bucket_location->US
DEBUG: ConfigParser: cloudfront_host->cloudfront.amazonaws.com
DEBUG: ConfigParser: default_mime_type->binary/octet-stream
DEBUG: ConfigParser: delete_removed->False
DEBUG: ConfigParser: dry_run->False
DEBUG: ConfigParser: enable_multipart->True
DEBUG: ConfigParser: encoding->UTF-8
DEBUG: ConfigParser: encrypt->False
DEBUG: ConfigParser: follow_symlinks->False
DEBUG: ConfigParser: force->False
DEBUG: ConfigParser: get_continue->False
DEBUG: ConfigParser: gpg_command->/usr/bin/gpg
DEBUG: ConfigParser: gpg_decrypt->%(gpg_command)s -d --verbose --no-use-agent --batch --yes --passphrase-fd %(passphrase_fd)s -o %(output_file)s %(input_file)s
DEBUG: ConfigParser: gpg_encrypt->%(gpg_command)s -c --verbose --no-use-agent --batch --yes --passphrase-fd %(passphrase_fd)s -o %(output_file)s %(input_file)s
DEBUG: ConfigParser: gpg_passphrase->...-3_chars...
DEBUG: ConfigParser: guess_mime_type->True
DEBUG: ConfigParser: host_base->nprs3.com
DEBUG: ConfigParser: host_bucket->%(bucket)s.nprs3.com
DEBUG: ConfigParser: human_readable_sizes->False
DEBUG: ConfigParser: invalidate_on_cf->False
DEBUG: ConfigParser: list_md5->False
DEBUG: ConfigParser: log_target_prefix->
DEBUG: ConfigParser: mime_type->
DEBUG: ConfigParser: multipart_chunk_size_mb->15
DEBUG: ConfigParser: preserve_attrs->True
DEBUG: ConfigParser: progress_meter->True
DEBUG: ConfigParser: proxy_host->
DEBUG: ConfigParser: proxy_port->0
DEBUG: ConfigParser: recursive->False
DEBUG: ConfigParser: recv_chunk->4096
DEBUG: ConfigParser: reduced_redundancy->False
DEBUG: ConfigParser: secret_key->J7...37_chars...Y
DEBUG: ConfigParser: send_chunk->4096
DEBUG: ConfigParser: simpledb_host->sdb.amazonaws.com
DEBUG: ConfigParser: skip_existing->False
DEBUG: ConfigParser: socket_timeout->300
DEBUG: ConfigParser: urlencoding_mode->normal
DEBUG: ConfigParser: use_https->False
DEBUG: ConfigParser: verbosity->WARNING
DEBUG: ConfigParser: website_endpoint->http://%(bucket)s.s3-website-%(location)s.amazonaws.com/
DEBUG: ConfigParser: website_error->
DEBUG: ConfigParser: website_index->index.html
DEBUG: Updating Config.Config encoding -> UTF-8
DEBUG: Updating Config.Config follow_symlinks -> False
DEBUG: Updating Config.Config verbosity -> 10
DEBUG: Unicodising 'put' using UTF-8
DEBUG: Unicodising '2' using UTF-8
DEBUG: Unicodising '20130718_ascvid_cheyennemizeTEST4.mov' using UTF-8
DEBUG: Unicodising 's3://linux' using UTF-8
DEBUG: Command: put
INFO: Compiling list of local files...
DEBUG: DeUnicodising u'' using UTF-8
DEBUG: DeUnicodising u'2' using UTF-8
INFO: Compiling list of local files...
DEBUG: DeUnicodising u'' using UTF-8
DEBUG: DeUnicodising u'20130718_ascvid_cheyennemizeTEST4.mov' using UTF-8
DEBUG: Unicodising '20130718_ascvid_cheyennemizeTEST4.mov' using UTF-8
DEBUG: Unicodising '20130718_ascvid_cheyennemizeTEST4.mov' using UTF-8
INFO: Applying --exclude/--include
DEBUG: CHECK: 20130718_ascvid_cheyennemizeTEST4.mov
DEBUG: PASS: 20130718_ascvid_cheyennemizeTEST4.mov
INFO: Summary: 1 local files to upload
DEBUG: Unicodising '20130718_ascvid_cheyennemizeTEST4.mov' using UTF-8
WARNING: File can not be uploaded: 20130718_ascvid_cheyennemizeTEST4.mov: Permission denied

We will need a command line tool to use in cronjobs, etc...so I am hoping I can this working soon.

Thanks agin for the help already.

Shain


Shain Miley | Manager of Systems and Infrastructure, Digital Media | smiley@xxxxxxx | 202.513.3649

________________________________________
From: Derek Yarnell [derek@xxxxxxxxxxxxxx]
Sent: Saturday, October 26, 2013 6:46 PM
To: Shain Miley; ceph-users@xxxxxxxx
Subject: Re:  Radosgw and large files

Hi Shain,

Yes we have tested and have working S3 Multipart support for files >5GB
(RHEL64/0.67.4).

However, crossftp unless you have pro it would seem does not support
multipart.  Dragondisk gives the error that I have seen when using a PUT
and not multipart, EntityTooLarge.  My guess is that it is not doing
Multipart.  Now from s3cmd's source it does support multipart have you
tried to give it a --debug 2 which will output boto's debug output I
believe.   That should tell you more about if it is correctly doing a
multipart upload.

Thanks,
derek

On 10/26/13, 5:10 PM, Shain Miley wrote:
> Hi,
>
> I am wondering if anyone has successfully been able to upload files
> larger than 5GB using radosgw.
>
> I have tried using various clients, including dragondisk, crossftp,
> s3cmd, etc...all of them have failed with a 'permission denied' response.
>
> Each of the clients say they support multi-part (and it appears that
> they do as files larger than 15MB gets split into multiple pieces)
> however I can only upload files smaller than 5GB up to this point.
>
> I am using 0.67.4 and Ubuntu 12.04.
>
> Thanks in advance,
>
> Shain
>
> Shain Miley | Manager of Systems and Infrastructure, Digital Media |
> smiley@xxxxxxx | 202.513.3649
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>


--
---
Derek T. Yarnell
University of Maryland
Institute for Advanced Computer Studies


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux