I'm am running a Ceph Cluster on 5 Servers, all with a single osd and acting as a client (kernel) for nearly half a year now and didn't encounter a lockup yet. Total storage is 3.25TB with about 600GB raw storage used, if that matters.
Dan van der Ster <dan@xxxxxxxxxxxxxx> schrieb am Di., 23. Apr. 2019, 09:33:
_______________________________________________On Mon, 22 Apr 2019, 22:20 Gregory Farnum, <gfarnum@xxxxxxxxxx> wrote:On Sat, Apr 20, 2019 at 9:29 AM Igor Podlesny <ceph-user@xxxxxxxx> wrote:
>
> I remember seeing reports in regards but it's being a while now.
> Can anyone tell?
No, this hasn't changed. It's unlikely it ever will; I think NFS
resolved the issue but it took a lot of ridiculous workarounds and
imposes a permanent memory cost on the client.On the other hand, we've been running osds and local kernel mounts through some ior stress testing and managed to lock up only one node, only once (and that was with a 2TB shared output file).Maybe the necessary memory pressure conditions get less likely as the number of clients and osds gets larger? (i.e. it's probably easy to trigger with one single node/osd because all IO is local, but for large clusters most IO is remote)... Dan-Greg
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com