Re: Data not distributed according to weights

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello Christian,
hello everyone,

> That would be my guess, as a similar 4 node cluster here with a flat, host
> only topology is having its data distributed as expected by the weights.

To verify this, I eliminated the room buckets.
Now all hosts are located within the root bucket.
(Same tree as the one below, just without room buckets)
However, after rebalancing, the data is still not distributed according to the weights.
(Almost the same as the one below, le10970 ~20%, all others ~80+-6%) 

Any further ideas what I am doing wrong? 

Thanks, Frank

> > In my rather heterogeneous setup ...
> >
> > -1      54.36   root default
> > -2      42.44           room 2.162
> > -4      6.09                    host le09091
> > 3       2.03                            osd.3   up      1
> > 1       2.03                            osd.1   up      1
> > 9       2.03                            osd.9   up      1
> > -6      36.35                   host le10970
> > 4       7.27                            osd.4   up      1
> > 5       7.27                            osd.5   up      1
> > 6       7.27                            osd.6   up      1
> > 7       7.27                            osd.7   up      1
> > 8       7.27                            osd.8   up      1
> > -3      11.92           room 2.166
> > -5      5.83                    host le09086
> > 2       2.03                            osd.2   up      1
> > 0       2.03                            osd.0   up      1
> > 10      1.77                            osd.10  up      1
> > -7      6.09                    host le08544
> > 11      2.03                            osd.11  up      1
> > 12      2.03                            osd.12  up      1
> > 13      2.03                            osd.13  up      1
> >
> > ... using size =3 for all pools,
> > the OSDs are not filled according to the weights
> > (which correspond to the disk sizes in TB).
> >
> > le09086 (osd)
> > /dev/sdb1    1.8T  1.5T  323G  83% /var/lib/ceph/osd/ceph-10
> > /dev/sdc1    2.1T  1.7T  408G  81% /var/lib/ceph/osd/ceph-0
> > /dev/sdd1    2.1T  1.7T  344G  84% /var/lib/ceph/osd/ceph-2
> > le09091 (osd)
> > /dev/sda1    2.1T  1.6T  447G  79% /var/lib/ceph/osd/ceph-9
> > /dev/sdc1    2.1T  1.8T  317G  85% /var/lib/ceph/osd/ceph-3
> > /dev/sdb1    2.1T  1.7T  384G  82% /var/lib/ceph/osd/ceph-1
> > le10970 (osd)
> > /dev/sdd1    7.3T  1.4T  5.9T  19% /var/lib/ceph/osd/ceph-6
> > /dev/sdf1     7.3T  1.5T  5.9T  21% /var/lib/ceph/osd/ceph-8
> > /dev/sde1     7.3T  1.6T  5.7T  22% /var/lib/ceph/osd/ceph-7
> > /dev/sdc1     7.3T  1.4T  6.0T  19% /var/lib/ceph/osd/ceph-5
> > /dev/sdb1     7.3T  1.5T  5.8T  21% /var/lib/ceph/osd/ceph-4
> > le08544 (osd)
> > /dev/sdc1     2.1T  1.6T  443G  79% /var/lib/ceph/osd/ceph-13
> > /dev/sdb1     2.1T  1.7T  339G  84% /var/lib/ceph/osd/ceph-12
> > /dev/sda1     2.1T  1.7T  375G  82% /var/lib/ceph/osd/ceph-11
> >
> > Clearly, I would like le10970 to be selected more often!
> >
> > Increasing pg(p)_num from 256 to 512 of all the pools
> > didn't help.
> >
> > Optimal tunables are used ...
> >
> > tunable choose_local_tries 0
> > tunable choose_local_fallback_tries 0
> > tunable choose_total_tries 50
> > tunable chooseleaf_descend_once 1
> > tunable chooseleaf_vary_r 1
> >
> > ... as well as the default algo (straw) and hash (0)
> > for all buckets. The ruleset is pretty much standard,
> > too ...
> >
> > rule replicated_ruleset {
> >         ruleset 0
> >         type replicated
> >         min_size 1
> >         max_size 10
> >         step take default
> >         step chooseleaf firstn 0 type host
> >         step emit
> > }
> >
> > ... so I would assume it should select 3 hosts
> > (originating from root) according to the weights.
> >
> > It is ceph version 0.87.
> >
> > Is it the room bucket, which circumvents
> > a distribution according tot he weights?
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux