Re: rsync kernel client cepfs mkstemp no space left on device

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Am 10.10.2016 um 10:05 schrieb Hauke Homburg:
> Am 07.10.2016 um 17:37 schrieb Gregory Farnum:
>> On Fri, Oct 7, 2016 at 7:15 AM, Hauke Homburg <hhomburg@xxxxxxxxxxxxxx> wrote:
>>> Hello,
>>>
>>> I have a Ceph Cluster with 5 Server, and 40 OSD. Aktual on this Cluster
>>> are 85GB Free Space, and the rsync dir has lots of Pictures and a Data
>>> Volume of 40GB.
>>>
>>> The Linux is a Centos 7 and the Last stable Ceph. The Client is a Debian
>>> 8 with Kernel 4 and the Cluster is with cephfs mounted.
>>>
>>> When i sync the Directory i see often the Message rsync mkstemp no space
>>> left on device (28). At this Point i can touch a File in anotherDiretory
>>> in the Cluster. In the Diretory i have ~ 630000 Files. Are this too much
>>> Files?
>> Yes, in recent releases CephFS limits you to 100k dentries in a single
>> directory fragment. This *includes* the "stray" directories that files
>> get moved into when you unlink them, and is intended to prevent issues
>> with very large folders. It will stop being a problem once we enable
>> automatic fragmenting (soon, hopefully).
>> You can change that by changing the "mds bal fragment size max"
>> config, but you're probably better off by figuring out if you've got
>> an over-large directory or if you're deleting files faster than the
>> cluster can keep up. There was a thread about this very recently and
>> John included some details about tuning if you check the archives. :)
>> -Greg
> Hello,
>
> Thanks for the answer.
> I enabled on the Cluster the mds bal frag = true Options.
>
> Today i read that i have to enable this option on the Client, too. With
> a Fuse mount i can do it with the ceph Binary. I use the Kernel Module.
> How can i do it there?
>
> Regards
>
> Hauke
>

Hello,

After some Discussion in our Team we have deleted die Cephfs and
switched to rados with ext4.

Now we want to realive the Setup:

1 Ceph Cluster Jewel 10.0.2.3
5 Server with Ceph 10.0.2.3 Client with rados installed. We map all 5
Mons of our Cluster into every rbd map call. To have a Failover.

Aktually we have the Problem, that we can store Data into The Cluster
with rsync, but when rsync is deleting Files, ext4 becomes Filesystem
errors.

I undersstand Ceph with rbd so, that i can use Ceph als Cluster
Filesystem like ocfs2. So i don't unterstand, why i have Filesytem Errors.

I read in some Postings here, Ceph needs Filesystem Locking like DLM. Is
this true? In the aktual Version Jewel? Doesn't do libceph this Locking?

Thanks for Help

Hauke

-- 
www.w3-creative.de

www.westchat.de


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux