Re: 3x replicated rbd pool ssd data spread across 4 osd's

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



So that changes the question to: why is ceph not distributing the pg's 
evenly across four osd's? 

[@c01 ~]# ceph osd df |egrep '^19|^20|^21|^30'
19   ssd 0.48000  1.00000  447G   133G   313G 29.81 0.70  16
20   ssd 0.48000  1.00000  447G   158G   288G 35.40 0.83  19
21   ssd 0.48000  1.00000  447G   208G   238G 46.67 1.10  20
30   ssd 0.48000  1.00000  447G   149G   297G 33.50 0.79  17

rbd.ssd: pg_num 8 pgp_num 8 

I will look into the balancer, but I am still curious why these 8 pg 
(8x8=64? + 8? = 72) are still not spread evenly. Why not 18 on every 
osd?

-----Original Message-----
From: Jack [mailto:ceph@xxxxxxxxxxxxxx] 
Sent: zondag 2 september 2018 14:06
To: ceph-users@xxxxxxxxxxxxxx
Subject: Re:  3x replicated rbd pool ssd data spread across 
4 osd's

ceph osd df will get you more information: variation & pg number for 
each OSD

Ceph does not spread object on a per-object basis, but on a pg-basis

The data repartition is thus not perfect You may increase your pg_num, 
and/or use the mgr balancer module
(http://docs.ceph.com/docs/mimic/mgr/balancer/)


On 09/02/2018 01:28 PM, Marc Roos wrote:
> 
> If I have only one rbd ssd pool, 3 replicated, and 4 ssd osd's. Why 
> are these objects so unevenly spread across the four osd's? Should 
> they all not have 162G?
> 
> 
> [@c01 ]# ceph osd status 2>&1
> +----+------+-------+-------+--------+---------+--------+---------+---
> +----+------+-------+-------+--------+---------+--------+---------+--
> ------+
> | id | host |  used | avail | wr ops | wr data | rd ops | rd data |   
> state   |
> +----+------+-------+-------+--------+---------+--------+---------+---
> +----+------+-------+-------+--------+---------+--------+---------+--
> ------+
> | 19 | c01  |  133G |  313G |    0   |     0   |    0   |     0   | 
> exists,up |
> | 20 | c02  |  158G |  288G |    0   |     0   |    0   |     0   | 
> exists,up |
> | 21 | c03  |  208G |  238G |    0   |     0   |    0   |     0   | 
> exists,up |
> | 30 | c04  |  149G |  297G |    0   |     0   |    0   |     0   | 
> exists,up |
> +----+------+-------+-------+--------+---------+--------+---------+---
> +----+------+-------+-------+--------+---------+--------+---------+--
> ------+
> 
> All objects in the rbd pool are 4MB not? Should be easy to spread them 

> evenly, what am I missing here?
> 
> 
> 
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux