Re: Problem balancing disk space

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Data balancing is depended on the following items:

1. osd counts, more is better, while 20 may be not enough.
2. PG counts, more is better. you can try to create a new pool with
more PGs than default.
3. Object size, little is better. More objects with little size is
better than less objects with big size.
4. Bucket type in the CrushMap. different bucket type means different
balance algorithm.
5. Maybe other issues.

You can use pg dump command to get details about object counts for
each PG, this may be helpful to locate your problem.

On 20 August 2014 04:43, Alphe Salas <asalas@xxxxxxxxx> wrote:
> Hello,
> for some reasons the balancing of disk space use on OSD is not properly
> working. Can you please give me hint to solve that issue?
>
> it is supposed that the proper difference betwin min and max osd disk space
> use, should be around 20%.
>
> Actually I see that it is more that 40%
>
> follow you will find the list of disk and real use
>
> osd10: Filesystem      Size  Used Avail Use% Mounted on
> osd10: /dev/sda1       1.8T  1.1T  646G  63% /var/lib/ceph/osd/ceph-18
> osd10: Filesystem      Size  Used Avail Use% Mounted on
> osd10: /dev/sdb1       1.8T  1.6T  109G  94% /var/lib/ceph/osd/ceph-19
> osd09: Filesystem      Size  Used Avail Use% Mounted on
> osd09: /dev/sda1       1.8T  1.5T  216G  88% /var/lib/ceph/osd/ceph-16
> osd09: Filesystem      Size  Used Avail Use% Mounted on
> osd09: /dev/sdb1       1.8T  895G  842G  52% /var/lib/ceph/osd/ceph-17
> osd08: Filesystem      Size  Used Avail Use% Mounted on
> osd08: /dev/sda1       1.8T  1.6T  153G  92% /var/lib/ceph/osd/ceph-10
> osd08: Filesystem      Size  Used Avail Use% Mounted on
> osd08: /dev/sdb1       1.8T  1.7T   84G  96% /var/lib/ceph/osd/ceph-11
> osd07: Filesystem      Size  Used Avail Use% Mounted on
> osd07: /dev/sda1       1.8T  1.5T  297G  83% /var/lib/ceph/osd/ceph-14
> osd07: Filesystem      Size  Used Avail Use% Mounted on
> osd07: /dev/sdb1       1.8T  1.5T  268G  85% /var/lib/ceph/osd/ceph-15
> osd06: Filesystem      Size  Used Avail Use% Mounted on
> osd06: /dev/sda1       1.8T  1.6T  193G  89% /var/lib/ceph/osd/ceph-12
> osd06: Filesystem      Size  Used Avail Use% Mounted on
> osd06: /dev/sdb1       1.8T  1.4T  305G  83% /var/lib/ceph/osd/ceph-13
> osd05: Filesystem      Size  Used Avail Use% Mounted on
> osd05: /dev/sda1       1.8T  1.3T  434G  76% /var/lib/ceph/osd/ceph-8
> osd05: Filesystem      Size  Used Avail Use% Mounted on
> osd05: /dev/sdb1       1.8T  1.2T  526G  70% /var/lib/ceph/osd/ceph-9
> osd04: Filesystem      Size  Used Avail Use% Mounted on
> osd04: /dev/sda1       1.8T  1.6T  169G  91% /var/lib/ceph/osd/ceph-6
> osd04: Filesystem      Size  Used Avail Use% Mounted on
> osd04: /dev/sdb1       1.8T  1.4T  313G  82% /var/lib/ceph/osd/ceph-7
> osd03: Filesystem      Size  Used Avail Use% Mounted on
> osd03: /dev/sda1       1.8T  1.6T  195G  89% /var/lib/ceph/osd/ceph-4
> osd03: Filesystem      Size  Used Avail Use% Mounted on
> osd03: /dev/sdb1       1.8T  1.3T  425G  76% /var/lib/ceph/osd/ceph-5
> osd02: Filesystem      Size  Used Avail Use% Mounted on
> osd02: /dev/sda1       1.8T  1.4T  362G  80% /var/lib/ceph/osd/ceph-2
> osd02: Filesystem      Size  Used Avail Use% Mounted on
> osd02: /dev/sdb1       1.8T  1.5T  211G  88% /var/lib/ceph/osd/ceph-3
> osd01: Filesystem      Size  Used Avail Use% Mounted on
> osd01: /dev/sda1       1.8T  1.4T  304G  83% /var/lib/ceph/osd/ceph-0
> osd01: Filesystem      Size  Used Avail Use% Mounted on
> osd01: /dev/sdb1       1.8T  1.3T  456G  74% /var/lib/ceph/osd/ceph-1
>
> ceph health detail gives an inacurate estimation of the problem (the
> percentages are wrong..)
>
> ceph health detail
> HEALTH_WARN 4 near full osd(s)
> osd.6 is near full at 85% (real 91%)
> osd.10 is near full at 86% (real 92%)
> osd.11 is near full at 90% (real 96%)
> osd.19 is near full at 88% (real 94%)
>
>
> Then as you can see on the above disk space use dump I get a usage span from
> 52% to 96%.
>
> The question is how can I force the re balancing. I tryed with ceph osd
> reweight-by-use 108 and still there is this amazing gap.
>
> actual osd tree is
>
> # id    weight  type name       up/down reweight
> -1      35.8    root default
> -2      3.58            host osd01
> 0       1.79                    osd.0   up      1
> 1       1.79                    osd.1   up      0.808
> -3      3.58            host osd02
> 2       1.79                    osd.2   up      0.9193
> 3       1.79                    osd.3   up      1
> -4      3.58            host osd03
> 4       1.79                    osd.4   up      1
> 5       1.79                    osd.5   up      1
> -5      3.58            host osd04
> 6       1.79                    osd.6   up      1
> 7       1.79                    osd.7   up      1
> -6      3.58            host osd05
> 8       1.79                    osd.8   up      0.7892
> 9       1.79                    osd.9   up      0.7458
> -7      3.58            host osd08
> 10      1.79                    osd.10  up      1
> 11      1.79                    osd.11  up      1
> -8      3.58            host osd06
> 12      1.79                    osd.12  up      1
> 13      1.79                    osd.13  up      1
> -9      3.58            host osd07
> 14      1.79                    osd.14  up      1
> 15      1.79                    osd.15  up      1
> -10     3.58            host osd09
> 16      1.79                    osd.16  up      1
> 17      1.79                    osd.17  up      1
> -11     3.58            host osd10
> 18      1.79                    osd.18  up      1
> 19      1.79                    osd.19  up      1
>
> Regards,
>
>
> --
> Alphe Salas
> I.T ingeneer
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html



-- 
Dong Yuan
Email:yuandong1222@xxxxxxxxx
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux