osd' balancing question

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello cephers
We're running firefly  ( 9.2.1 ) 
I'm trying to re balance  our cluster's osd and from some reason it looks like the re balance is going the wrong way :
What's i'm trying to do is to reduce the loads from osd-14  ( ceph osd crush reweight osd.14 0.75 ) but what i see is the the backfill process  is moving pgs to osd-29 which is also  86% full 
i wonder why the crash doesn't map to the less occupied  osd-s  (  3 , 4  6 for example )
 Any input is much appreciated . 



2017-01-03 05:59:20.877705 7f3e6a0d6700  0 log_channel(cluster) log [INF] : 2.2cb starting backfill to osd.29 from (0'0,0'0] MAX to 131306'8029954
2017-01-03 05:59:20.877841 7f3e670d0700  0 log_channel(cluster) log [INF] : 2.30d starting backfill to osd.10 from (0'0,0'0] MAX to 131306'8721158
2017-01-03 05:59:31.374323 7f3e356b0700  0 -- 10.63.4.3:6826/3125306 >> 10.63.4.5:6821/3162046 pipe(0x7f3e9d513000 sd=322 :6826 s=0 pgs=0 cs=0 l=0 c=0x7f3ea72b5de0).accept connect_seq 1605 vs existing 1605 state standby
2017-01-03 05:59:31.374440 7f3e356b0700  0 -- 10.63.4.3:6826/3125306 >> 10.63.4.5:6821/3162046 pipe(0x7f3e9d513000 sd=322 :6826 s=0 pgs=0 cs=0 l=0 c=0x7f3ea72b5de0).accept connect_seq 1606 vs existing 1605 state standby
^C
root@ecprdbcph03-opens:/var/log/ceph# df -h
Filesystem                           Size  Used Avail Use% Mounted on
udev                                  32G  4.0K   32G   1% /dev
tmpfs                                6.3G  1.4M  6.3G   1% /run
/dev/dm-1                            106G  4.1G   96G   5% /
none                                 4.0K     0  4.0K   0% /sys/fs/cgroup
none                                 5.0M     0  5.0M   0% /run/lock
none                                  32G     0   32G   0% /run/shm
none                                 100M     0  100M   0% /run/user
/dev/sdk2                            465M   50M  391M  12% /boot
/dev/sdk1                            512M  3.4M  509M   1% /boot/efi
ec-mapr-prd:/mapr/ec-mapr-prd/homes  262T  143T  119T  55% /export/home
/dev/sde1                            889G  640G  250G  72% /var/lib/ceph/osd/ceph-3
/dev/sdf1                            889G  656G  234G  74% /var/lib/ceph/osd/ceph-4
/dev/sdg1                            889G  583G  307G  66% /var/lib/ceph/osd/ceph-6
/dev/sda1                            889G  559G  331G  63% /var/lib/ceph/osd/ceph-8
/dev/sdb1                            889G  651G  239G  74% /var/lib/ceph/osd/ceph-10
/dev/sdc1                            889G  751G  139G  85% /var/lib/ceph/osd/ceph-12
/dev/sdh1                            889G  759G  131G  86% /var/lib/ceph/osd/ceph-14
/dev/sdi1                            889G  763G  127G  86% /var/lib/ceph/osd/ceph-16
/dev/sdj1                            889G  732G  158G  83% /var/lib/ceph/osd/ceph-18
/dev/sdd1                            889G  756G  134G  86% /var/lib/ceph/osd/ceph-29
root@ecprdbcph03-opens:/var/log/ceph#
 
Thanks



Yair Magnezi
Storage & Data Protection TL   // Kenshoo




This e-mail, as well as any attached document, may contain material which is confidential and privileged and may include trademark, copyright and other intellectual property rights that are proprietary to Kenshoo Ltd,  its subsidiaries or affiliates ("Kenshoo"). This e-mail and its attachments may be read, copied and used only by the addressee for the purpose(s) for which it was disclosed herein. If you have received it in error, please destroy the message and any attachment, and contact us immediately. If you are not the intended recipient, be aware that any review, reliance, disclosure, copying, distribution or use of the contents of this message without Kenshoo's express permission is strictly prohibited.
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux