Thanks, I moved back to crush-compat mapping, the pool that was at "90% full" is now under 76% full. Before doing that, I had the automatic balancer off, and ran 'ceph balancer optimize test'. It ran for 12 hours before I killed it. In upmap mode, it was "balanced" or at least as balanced as it could get. crush-compat seems to be much more flexible, and more useful in my situation. -- Adam On Wed, Sep 11, 2019 at 10:04 PM Konstantin Shalygin <k0ste@xxxxxxxx> wrote: > > We're using Nautilus 14.2.2 (upgrading soon to 14.2.3) on 29 CentOS osd servers. > > We've got a large variation of disk sizes and host densities. Such > that the default crush mappings lead to an unbalanced data and pg > distribution. > > We enabled the balancer manager module in pg upmap mode. The balancer > commands frequently hang indefinitely when enabled and then queried. > Even issuing a balancer off will hang for hours unless issued within > about a minute of the manager restarting. I digress. > > In upmap mode, it looks like ceph only moves osd mappings within a > host. Is this the case? > > I bring this up because we've got one disk that is sitting at 88% > utilization and I've been unable to bring this down. The next most > utilized disks are at 80%, and even then, I think that could be > reduced. > > If the limitation is that upmap mode cannot map to osds to different > hosts, than that might be something to document. As it is a > significant difference to crush-compat. > > Another thing to document would be how to move between the two modes. > > I think this is needed to move between crush-compat and upmap: ceph > osd crush weight-set rm-compat > > I don't know about the reverse, though. > > ceph osd df tree [1] > pg upmap items from the osdmap [2] > > [1] https://people.cs.ksu.edu/~mozes/ceph_balancer_query/ceph_osd_df_tree.txt > [2] https://people.cs.ksu.edu/~mozes/ceph_balancer_query/pg_upmap_items.txt > > To remove upmaps you can execute `ceph osd rm-pg-upmap-items ${upmap}` from your dump. > > Don't forget to "off" balancer before that operation. > > > > k _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com