Re: dealing with the full osd / help reweight

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Yeah,I think the main reason is the setting of pg_num and pgp_num of some key pool.
This site will tell you the correct value:http://ceph.com/pgcalc/


Before you adjust pg_num and pgp_num,if this is a product environment,you should set as Christian Balzer said:
--- 
osd_max_backfills = 1 
osd_backfill_scan_min = 4 
osd_backfill_scan_max = 32 
osd_recovery_max_active = 1 
osd recovery threads = 1 
osd recovery op priority = 1 
—
you can use 'ceph tell osd.* injectargs “--odd-max-backfills 1” to change this setting.

and recommended process is 8->16->32->64->128->256->…  the more you increase the more data will rebalance,and more time it will take.so be careful.


--  
hnuzhoulin2@xxxxxxxxx

在 2016年3月24日 20:57:17, Jacek Jarosiewicz (jjarosiewicz@xxxxxxxxxxxxx) 写到:
> [root@cf01 ceph]# ceph osd pool ls detail
> pool 0 'vms' replicated size 2 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 64 pgp_num 64 last_change 67 flags hashpspool stripe_width 0
> pool 1 '.rgw.root' replicated size 2 min_size 1 crush_ruleset 0
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 117 flags hashpspool
> stripe_width 0
> pool 2 '.rgw.control' replicated size 2 min_size 1 crush_ruleset 0
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 118 flags hashpspool
> stripe_width 0
> pool 3 '.rgw.gc' replicated size 2 min_size 1 crush_ruleset 0
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 119 flags hashpspool
> stripe_width 0
> pool 4 '.rgw.buckets_cache' replicated size 2 min_size 1 crush_ruleset 0
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 121 flags hashpspool
> stripe_width 0
> pool 5 '.rgw.buckets.index' replicated size 2 min_size 1 crush_ruleset 0
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 122 flags hashpspool
> stripe_width 0
> pool 6 '.rgw.buckets.extra' replicated size 2 min_size 1 crush_ruleset 0
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 123 flags hashpspool
> stripe_width 0
> pool 7 '.log' replicated size 2 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 8 pgp_num 8 last_change 124 flags hashpspool stripe_width 0
> pool 8 '.intent-log' replicated size 2 min_size 1 crush_ruleset 0
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 125 flags hashpspool
> stripe_width 0
> pool 9 '.usage' replicated size 2 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 8 pgp_num 8 last_change 126 flags hashpspool stripe_width 0
> pool 10 '.users' replicated size 2 min_size 1 crush_ruleset 0
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 127 flags hashpspool
> stripe_width 0
> pool 11 '.users.email' replicated size 2 min_size 1 crush_ruleset 0
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 128 flags hashpspool
> stripe_width 0
> pool 12 '.users.swift' replicated size 2 min_size 1 crush_ruleset 0
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 129 flags hashpspool
> stripe_width 0
> pool 13 '.users.uid' replicated size 2 min_size 1 crush_ruleset 0
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 130 flags hashpspool
> stripe_width 0
> pool 14 '.rgw.buckets' replicated size 2 min_size 1 crush_ruleset 0
> object_hash rjenkins pg_num 64 pgp_num 64 last_change 5717 flags
> hashpspool stripe_width 0
> pool 15 '.rgw' replicated size 2 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 8 pgp_num 8 last_change 135 owner 18446744073709551615
> flags hashpspool stripe_width 0
> pool 17 'one' replicated size 3 min_size 2 crush_ruleset 0 object_hash
> rjenkins pg_num 64 pgp_num 64 last_change 611 flags hashpspool
> stripe_width 0
> removed_snaps [1~e,13~1]
>  
> I've tried reweight-by-utilization, but after some data shifting the
> cluster came up with a near full osd again..
>  
> Do I assume correctly that a lower weight of an osd means 'use that osd
> less'?
>  
> J
>  
> On 03/24/2016 01:43 PM, koukou73gr wrote:
> > What is your pool size? 304 pgs sound awfuly small for 20 OSDs.
> > More pgs will help distribute full pgs better.
> >
> > But with a full or near full OSD in hand, increasing pgs is a no-no
> > operation. If you search in the list archive, I believe there was a
> > thread last month or so which provided a walkthrough-sort of for dealing
> > with uneven distribution and a full OSD.
> >
> > -K.
> >
> >
> > On 03/24/2016 01:54 PM, Jacek Jarosiewicz wrote:
> >> disk usage on the full osd is as below. What are the *_TEMP directories
> >> for? How can I make sure which pg directories are safe to remove?
> >>
> >> [root@cf04 current]# du -hs *
> >> 156G 0.14_head
> >> 156G 0.21_head
> >> 155G 0.32_head
> >> 157G 0.3a_head
> >> 155G 0.e_head
> >> 156G 0.f_head
> >> 40K 10.2_head
> >> 4.0K 11.3_head
> >> 218G 14.13_head
> >> 218G 14.15_head
> >> 218G 14.1b_head
> >> 219G 14.1f_head
> >> 9.1G 14.29_head
> >> 219G 14.2a_head
> >> 75G 14.2d_head
> >> 125G 14.2e_head
> >> 113G 14.32_head
> >> 163G 14.33_head
> >> 218G 14.35_head
> >> 151G 14.39_head
> >> 218G 14.3b_head
> >> 103G 14.3d_head
> >> 217G 14.3f_head
> >> 219G 14.a_head
> >> 773M 17.0_head
> >> 814M 17.10_head
> >> 4.0K 17.10_TEMP
> >> 747M 17.19_head
> >> 4.0K 17.19_TEMP
> >> 669M 17.1b_head
> >> 659M 17.1c_head
> >> 638M 17.1f_head
> >> 681M 17.30_head
> >> 4.0K 17.30_TEMP
> >> 721M 17.34_head
> >> 695M 17.3d_head
> >> 726M 17.3e_head
> >> 734M 17.3f_head
> >> 4.0K 17.3f_TEMP
> >> 670M 17.d_head
> >> 597M 17.e_head
> >> 4.0K 17.e_TEMP
> >> 4.0K 1.7_head
> >> 34M 5.1_head
> >> 34M 5.6_head
> >> 4.0K 9.6_head
> >> 4.0K commit_op_seq
> >> 30M meta
> >> 0 nosnap
> >> 614M omap
> >>
> >>
> >>
> >> On 03/24/2016 10:11 AM, Jacek Jarosiewicz wrote:
> >>> Hi!
> >>>
> >>> I have a problem with the osds getting full on our cluster.
> >>>
> >>
> >
>  
>  
> --
> Jacek Jarosiewicz
> Administrator Systemów Informatycznych
>  
> ----------------------------------------------------------------------------------------  
> SUPERMEDIA Sp. z o.o. z siedzibą w Warszawie
> ul. Senatorska 13/15, 00-075 Warszawa
> Sąd Rejonowy dla m.st.Warszawy, XII Wydział Gospodarczy Krajowego
> Rejestru Sądowego,
> nr KRS 0000029537; kapitał zakładowy 44.556.000,00 zł
> NIP: 957-05-49-503
> Adres korespondencyjny: ul. Jubilerska 10, 04-190 Warszawa
>  
> ----------------------------------------------------------------------------------------  
> SUPERMEDIA -> http://www.supermedia.pl
> dostep do internetu - hosting - kolokacja - lacza - telefonia
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>  

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux