Thanks for your advice.
I will try to reweight osds of my cluster.
Why ceph is so sensitive to unblanced pg distribution during high load?
ceph osd df result is: https://pastebin.com/ur4Q9jsA.
ceph osd perf result is: https://pastebin.com/87DitPhV
There is no osd with very high pg count compare to others. When the
wirte test load is low everything seems fine, but during high write load test,
some of the osds with higher pg can have 3-10 time of fs_apply_latency compare
to others.
My guess is the high loaded osds kinda slowed the whole cluster(because I
have only one pool with all osds)to the level of how fast they can handle.
So other osd has lower load and have a good latency.
Is this expected during high load(Indicate the load is too hight
for current cluster to hanlde)?
How does luminous solve the unevenly pg distribution problem?I read about
there is a pg-upmap exception table in the osdmap in luminous 12.2.x. It is
said to use this it is possible to achive perfect pg distribution among
osds.
2018-03-09
shadow_lin
|
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com