ceph osd crush tunables optimal AND add new OSD at the same time

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Sage, Andrija & List

I have seen the tuneables issue on our cluster when I upgraded to firefly.

I ended up going back to legacy settings after about an hour as my cluster is of 55 3TB OSD?s over 5 nodes and it decided it needed to move around 32% of our data, which after an hour all of our vm?s were frozen and I had to revert the change back to legacy settings and wait about the same time again until our cluster had recovered and reboot our vms. (wasn?t really expecting that one from the patch notes)

Also our CPU usage went through the roof as well on our nodes, do you per chance have your metadata servers co-located on your osd nodes as we do?  I?ve been thinking about trying to move these to dedicated nodes as it may resolve our issues.

Regards,
Quenten

From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf Of Andrija Panic
Sent: Tuesday, 15 July 2014 8:38 PM
To: Sage Weil
Cc: ceph-users at lists.ceph.com
Subject: Re: ceph osd crush tunables optimal AND add new OSD at the same time

Hi Sage,

since this problem is tunables-related, do we need to expect same behavior or not  when we do regular data rebalancing caused by adding new/removing OSD? I guess not, but would like your confirmation.
I'm already on optimal tunables, but I'm afraid to test this by i.e. shuting down 1 OSD.

Thanks,
Andrija

On 14 July 2014 18:18, Sage Weil <sweil at redhat.com<mailto:sweil at redhat.com>> wrote:
I've added some additional notes/warnings to the upgrade and release
notes:

 https://github.com/ceph/ceph/commit/fc597e5e3473d7db6548405ce347ca7732832451

If there is somewhere else where you think a warning flag would be useful,
let me know!

Generally speaking, we want to be able to cope with huge data rebalances
without interrupting service.  It's an ongoing process of improving the
recovery vs client prioritization, though, and removing sources of
overhead related to rebalancing... and it's clearly not perfect yet. :/

sage


On Sun, 13 Jul 2014, Andrija Panic wrote:

> Hi,
> after seting ceph upgrade (0.72.2 to 0.80.3) I have issued "ceph osd crush
> tunables optimal" and after only few minutes I have added 2 more OSDs to the
> CEPH cluster...
>
> So these 2 changes were more or a less done at the same time - rebalancing
> because of tunables optimal, and rebalancing because of adding new OSD...
>
> Result - all VMs living on CEPH storage have gone mad, no disk access
> efectively, blocked so to speak.
>
> Since this rebalancing took 5h-6h, I had bunch of VMs down for that long...
>
> Did I do wrong by causing "2 rebalancing" to happen at the same time ?
> Is this behaviour normal, to cause great load on all VMs because they are
> unable to access CEPH storage efectively ?
>
> Thanks for any input...
> --
>
> Andrija Pani?
>
>



--

Andrija Pani?
--------------------------------------
  http://admintweets.com
--------------------------------------
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20140716/718a5983/attachment.htm>


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux