Udo, I had all VMs completely unoperational - so don't set "optimal" for now... On 14 July 2014 20:48, Udo Lembke <ulembke at polarzone.de> wrote: > Hi, > which values are all changed with "ceph osd crush tunables optimal"? > > Is it perhaps possible to change some parameter the weekends before the > upgrade is running, to have more time? > (depends if the parameter are available in 0.72...). > > The warning told, it's can take days... we have an cluster with 5 > storage node and 12 4TB-osd-disk each (60 osd), replica 2. The cluster > is 60% filled. > Networkconnection 10Gb. > Takes tunables optimal in such an configuration one, two or more days? > > Udo > > On 14.07.2014 18:18, Sage Weil wrote: > > I've added some additional notes/warnings to the upgrade and release > > notes: > > > > > https://github.com/ceph/ceph/commit/fc597e5e3473d7db6548405ce347ca7732832451 > > > > If there is somewhere else where you think a warning flag would be > useful, > > let me know! > > > > Generally speaking, we want to be able to cope with huge data rebalances > > without interrupting service. It's an ongoing process of improving the > > recovery vs client prioritization, though, and removing sources of > > overhead related to rebalancing... and it's clearly not perfect yet. :/ > > > > sage > > > > > > > > _______________________________________________ > ceph-users mailing list > ceph-users at lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > -- Andrija Pani? -------------------------------------- http://admintweets.com -------------------------------------- -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20140714/7a0f89b2/attachment.htm>