On Fri, 31 Aug 2012, Andrew Thompson wrote: > On 8/31/2012 7:11 AM, Xiaopong Tran wrote: > > Hi, > > > > Ceph storage on each disk in the cluster is very unbalanced. On each > > node, the data seems to go to one or two disks, while other disks > > are almost empty. > > > > I can't find anything wrong from the crush map, it's just the > > default for now. Attached is the crush map. > > Have you been reweight-ing osds? I went round and round with my cluster a few > days ago reloading different crush maps only to find that it re-injecting a > crush map didn't seem to overwrite reweights. > > Take a look at `ceph osd tree` to see if the reweight column matches the > weight column. Note that the ideal situation is for reweight to be 1, regardless of what the crush weight is. If you find the utilizations are skewed, I would look for other causes before resorting to reweight-by-utilization; it is meant to adjust the normal statistical variation you expect from a (pseudo)random placement, but if the variance is high there is likely another cause. sage > > Note: I'm new at this. This is my experience only, with 0.48.1, and may not be > entirely correct. > > -- > Andrew Thompson > http://aktzero.com/ > > -- > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html > > -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html