Re: poor data distribution

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,
I found something else what I think can help.
PG distribution it seems isn't ok.
Graph: http://dysk.onet.pl/link/AVzTe
All PGS is from 70 to 140 per OSD.
Primary 15 to 58 per OSD.

Is there some way to fix it?

--
Regards
Dominik


2014-01-30 Dominik Mostowiec <dominikmostowiec@xxxxxxxxx>:
> Hi,
> I found something else.
> 'ceph pg dump' shows PGs:
> - with zero or near zero objects count
> - with ~6,5k objects,  size ~1,4G
> - with ~13k objects, size ~2,8G
> This can be a reason of wrong data distribution on OSD's?
>
> ---
> Regards
> Dominik
>
>
> 2014-01-30 Dominik Mostowiec <dominikmostowiec@xxxxxxxxx>:
>> Hi,
>> I have problem with data distribution.
>> Smallest disk usage 40% vs highest 82%.
>> All PGS: 6504.
>> Almost all data is in '.rgw.buckets' pool with pg_num 4800.
>> The best way to better data distribution is increese pg_num in this pool?
>> Is thre another way? ( eg crush tunables, or something like that ..)
>>
>> Config: 9 hosts X 22 OSD = 198 OSDs
>> replica = 3
>>
>> ceph -v
>> ceph version 0.67.5 (a60ac9194718083a4b6a225fc17cad6096c69bd1)
>>
>> ceph osd dump:
>> pool 0 'data' rep size 3 min_size 1 crush_ruleset 0 object_hash
>> rjenkins pg_num 64 pgp_num 64 last_change 28459 owner 0
>> crash_replay_interval 45
>> pool 1 'metadata' rep size 3 min_size 1 crush_ruleset 1 object_hash
>> rjenkins pg_num 64 pgp_num 64 last_change 28460 owner 0
>> pool 2 'rbd' rep size 3 min_size 1 crush_ruleset 2 object_hash
>> rjenkins pg_num 64 pgp_num 64 last_change 28461 owner 0
>> pool 3 '.rgw.buckets' rep size 3 min_size 1 crush_ruleset 0
>> object_hash rjenkins pg_num 4800 pgp_num 4800 last_change 28462 owner
>> 0
>> pool 4 '.log' rep size 3 min_size 1 crush_ruleset 0 object_hash
>> rjenkins pg_num 1440 pgp_num 1440 last_change 28463 owner 0
>> pool 5 '.rgw' rep size 3 min_size 1 crush_ruleset 0 object_hash
>> rjenkins pg_num 8 pgp_num 8 last_change 28464 owner 0
>> pool 6 '.users.uid' rep size 3 min_size 1 crush_ruleset 0 object_hash
>> rjenkins pg_num 8 pgp_num 8 last_change 28465 owner 0
>> pool 7 '.users' rep size 3 min_size 1 crush_ruleset 0 object_hash
>> rjenkins pg_num 8 pgp_num 8 last_change 28466 owner 0
>> pool 8 '.usage' rep size 2 min_size 1 crush_ruleset 0 object_hash
>> rjenkins pg_num 8 pgp_num 8 last_change 28467 owner
>> 18446744073709551615
>> pool 9 '.intent-log' rep size 3 min_size 1 crush_ruleset 0 object_hash
>> rjenkins pg_num 8 pgp_num 8 last_change 28468 owner
>> 18446744073709551615
>> pool 10 '.rgw.control' rep size 3 min_size 1 crush_ruleset 0
>> object_hash rjenkins pg_num 8 pgp_num 8 last_change 33485 owner
>> 18446744073709551615
>> pool 11 '.rgw.gc' rep size 3 min_size 1 crush_ruleset 0 object_hash
>> rjenkins pg_num 8 pgp_num 8 last_change 33487 owner
>> 18446744073709551615
>> pool 12 '.rgw.root' rep size 2 min_size 1 crush_ruleset 0 object_hash
>> rjenkins pg_num 8 pgp_num 8 last_change 44540 owner 0
>> pool 13 '' rep size 2 min_size 1 crush_ruleset 0 object_hash rjenkins
>> pg_num 8 pgp_num 8 last_change 46912 owner 0
>>
>> rados df:
>> pool name       category                 KB      objects       clones
>>    degraded      unfound           rd        rd KB           wr
>> wr KB
>>                 -                          0            0            0
>>            0           0            0            0            0
>>     0
>> .intent-log     -                          0            0            0
>>            0           0            1            0      1893949
>> 1893697
>> .log            -                      69222         6587            0
>>            0           0            0            0    172911881
>> 172884792
>> .rgw            -                     114449       813564            0
>>            0           0     51094886     39834914      3606364
>> 1047115
>> .rgw.buckets    -                10371229682     54367877            0
>>            0           0   2610759046  54967451976   1930820947
>> 24031291524
>> ....
>> empty pools
>>
>> --
>> Regards
>> Dominik
>
>
>
> --
> Pozdrawiam
> Dominik



-- 
Pozdrawiam
Dominik
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux