Your EC profile requires 5 servers to be healthy. When you remove 1 OSD from the cluster, it recovers by moving all of the copies on that OSD to other OSDs in the same host. However when you remove an entire host, it cannot store 5 copies of the data on the 4 remaining servers with your crush rules. The EC profile you're using does not work with this type of testing based on your hardware configuration.
On Tue, Nov 28, 2017 at 8:43 AM Jakub Jaszewski <jaszewski.jakub@xxxxxxxxx> wrote:
_______________________________________________Hi, I'm trying to understand erasure coded pools and why CRUSH rules seem to work for only part of PGs in EC pools.Basically what I'm trying to do is to check erasure coded pool recovering behaviour after the single OSD or single HOST failure.I noticed that in case of HOST failure only part of PGs get recovered to active+remapped when other PGs remain in active+undersized+degraded state. Why??EC pool profile I use is k=3 , m=2.Also I'm not really sure what is the meaning of all steps of below crush rule (perhaps it is the root cause).rule ecpool_3_2 {ruleset 1type erasuremin_size 3max_size 5step set_chooseleaf_tries 5 # should I maybe try to increase this number of retry ?? Can I apply the changes to existing EC crush rule and pool or need to create a new one ?step set_choose_tries 100step take defaultstep chooseleaf indep 0 type host # Does it allow to choose more than one OSD from single HOST but first trying to get only one OSD per HOST if there are enough HOSTs in the cluster?step emit}ceph version 10.2.9 (jewel)# INITIAL CLUSTER STATEroot@host01:~# ceph osd treeID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY-1 218.18401 root default-6 218.18401 region MyRegion-5 218.18401 datacenter MyDC-4 218.18401 room MyRoom-3 43.63699 rack Rack01-2 43.63699 host host010 3.63599 osd.0 up 1.00000 1.000003 3.63599 osd.3 up 1.00000 1.000004 3.63599 osd.4 up 1.00000 1.000006 3.63599 osd.6 up 1.00000 1.000008 3.63599 osd.8 up 1.00000 1.0000010 3.63599 osd.10 up 1.00000 1.0000012 3.63599 osd.12 up 1.00000 1.0000014 3.63599 osd.14 up 1.00000 1.0000016 3.63599 osd.16 up 1.00000 1.0000019 3.63599 osd.19 up 1.00000 1.0000022 3.63599 osd.22 up 1.00000 1.0000025 3.63599 osd.25 up 1.00000 1.00000-8 43.63699 rack Rack02-7 43.63699 host host021 3.63599 osd.1 up 1.00000 1.000002 3.63599 osd.2 up 1.00000 1.000005 3.63599 osd.5 up 1.00000 1.000007 3.63599 osd.7 up 1.00000 1.000009 3.63599 osd.9 up 1.00000 1.0000011 3.63599 osd.11 up 1.00000 1.0000013 3.63599 osd.13 up 1.00000 1.0000015 3.63599 osd.15 up 1.00000 1.0000017 3.63599 osd.17 up 1.00000 1.0000020 3.63599 osd.20 up 1.00000 1.0000023 3.63599 osd.23 up 1.00000 1.0000026 3.63599 osd.26 up 1.00000 1.00000-10 130.91000 rack Rack03-9 43.63699 host host0318 3.63599 osd.18 up 1.00000 1.0000021 3.63599 osd.21 up 1.00000 1.0000024 3.63599 osd.24 up 1.00000 1.0000027 3.63599 osd.27 up 1.00000 1.0000028 3.63599 osd.28 up 1.00000 1.0000029 3.63599 osd.29 up 1.00000 1.0000030 3.63599 osd.30 up 1.00000 1.0000031 3.63599 osd.31 up 1.00000 1.0000032 3.63599 osd.32 up 1.00000 1.0000033 3.63599 osd.33 up 1.00000 1.0000034 3.63599 osd.34 up 1.00000 1.0000035 3.63599 osd.35 up 1.00000 1.00000-11 43.63699 host host0436 3.63599 osd.36 up 1.00000 1.0000037 3.63599 osd.37 up 1.00000 1.0000038 3.63599 osd.38 up 1.00000 1.0000039 3.63599 osd.39 up 1.00000 1.0000040 3.63599 osd.40 up 1.00000 1.0000041 3.63599 osd.41 up 1.00000 1.0000042 3.63599 osd.42 up 1.00000 1.0000043 3.63599 osd.43 up 1.00000 1.0000044 3.63599 osd.44 up 1.00000 1.0000045 3.63599 osd.45 up 1.00000 1.0000046 3.63599 osd.46 up 1.00000 1.0000047 3.63599 osd.47 up 1.00000 1.00000-12 43.63699 host host0548 3.63599 osd.48 up 1.00000 1.0000049 3.63599 osd.49 up 1.00000 1.0000050 3.63599 osd.50 up 1.00000 1.0000051 3.63599 osd.51 up 1.00000 1.0000052 3.63599 osd.52 up 1.00000 1.0000053 3.63599 osd.53 up 1.00000 1.0000054 3.63599 osd.54 up 1.00000 1.0000055 3.63599 osd.55 up 1.00000 1.0000056 3.63599 osd.56 up 1.00000 1.0000057 3.63599 osd.57 up 1.00000 1.0000058 3.63599 osd.58 up 1.00000 1.0000059 3.63599 osd.59 up 1.00000 1.00000root@host01:~# ceph -wcluster a6f73750-1972-47f6-bcf5-a99753be65adhealth HEALTH_OKmonmap e2: 3 mons at {host01=10.212.32.23:6789/0,host02=10.212.32.24:6789/0,host03=10.212.32.25:6789/0}election epoch 22, quorum 0,1,2 host01,host02,host03osdmap e527: 60 osds: 60 up, 60 inflags sortbitwise,require_jewel_osdspgmap v57164: 3736 pgs, 19 pools, 10343 bytes data, 241 objects4665 MB used, 218 TB / 218 TB avail3736 active+clean2017-11-28 07:38:52.350228 mon.0 [INF] pgmap v57163: 3736 pgs: 3736 active+clean; 10343 bytes data, 4665 MB used, 218 TB / 218 TB avail...root@host01:~#In the 1st scenario I stop single OSD (id 48, host host05) and after 5 minutes cluster start to recover by remapping PGs using other OSD from HOST host05.In the 2nd scenario, I stop all Ceph services on one HOST host05# FIND ALL PGs USING OSDs FROM HOST host05root@host01:~# ceph pg dump pgs_brief |egrep '\[48|,48|\[49|,49|\[50|,50|\[51|,51|\[52|,52|\[53|,53|\[54|,54|\[55|,55|\[56|,56|\[57|,57|\[58|,58|\[59|,59' > PGs_on_HOST_host05dumped pgs_brief in format plainroot@host01:~# wc -l PGs_on_HOST_host052556 PGs_on_HOST_host05# STOP ALL CEPH SERVICES on HOST host05root@host05:~# systemctl stop ceph.targetroot@host01:~# ceph -wcluster a6f73750-1972-47f6-bcf5-a99753be65adhealth HEALTH_OKmonmap e2: 3 mons at {host01=10.212.32.23:6789/0,host02=10.212.32.24:6789/0,host03=10.212.32.25:6789/0}election epoch 22, quorum 0,1,2 host01,host02,host03osdmap e538: 60 osds: 59 up, 59 inflags sortbitwise,require_jewel_osdspgmap v57405: 3736 pgs, 19 pools, 10343 bytes data, 241 objects4581 MB used, 214 TB / 214 TB avail3736 active+clean2017-11-28 08:08:21.349340 mon.0 [INF] pgmap v57405: 3736 pgs: 3736 active+clean; 10343 bytes data, 4581 MB used, 214 TB / 214 TB avail2017-11-28 08:08:33.082249 mon.0 [INF] osd.57 marked itself down2017-11-28 08:08:33.082607 mon.0 [INF] osd.49 marked itself down2017-11-28 08:08:33.082899 mon.0 [INF] osd.59 marked itself down2017-11-28 08:08:33.083471 mon.0 [INF] osd.56 marked itself down2017-11-28 08:08:33.084091 mon.0 [INF] osd.58 marked itself down2017-11-28 08:08:33.084842 mon.0 [INF] osd.53 marked itself down2017-11-28 08:08:33.085373 mon.0 [INF] osd.50 marked itself down2017-11-28 08:08:33.085830 mon.0 [INF] osd.54 marked itself down2017-11-28 08:08:33.086437 mon.0 [INF] osd.55 marked itself down2017-11-28 08:08:33.086664 mon.0 [INF] osd.52 marked itself down2017-11-28 08:08:33.086970 mon.0 [INF] osd.51 marked itself down2017-11-28 08:08:33.246299 mon.0 [INF] osdmap e539: 60 osds: 48 up, 59 in2017-11-28 08:08:33.253694 mon.0 [INF] pgmap v57406: 3736 pgs: 3736 active+clean; 10343 bytes data, 4581 MB used, 214 TB / 214 TB avail2017-11-28 08:08:34.333012 mon.0 [INF] osdmap e540: 60 osds: 48 up, 59 in2017-11-28 08:08:34.348753 mon.0 [INF] pgmap v57407: 3736 pgs: 64 peering, 658 stale+active+clean, 3014 active+clean; 10343 bytes data, 4581 MB used, 214 TB / 214 TB avail2017-11-28 08:08:35.344372 mon.0 [INF] pgmap v57408: 3736 pgs: 4 active+undersized+degraded, 42 activating+undersized+degraded, 64 peering, 648 stale+active+clean, 2978 active+clean; 10343 bytes data, 4581 MB used, 214 TB / 214 TB avail2017-11-28 08:08:36.375645 mon.0 [INF] pgmap v57409: 3736 pgs: 268 active+undersized+degraded, 42 activating+undersized+degraded, 64 peering, 578 stale+active+clean, 2784 active+clean; 10343 bytes data, 4584 MB used, 214 TB / 214 TB avail; 24/791 objects degraded (3.034%)2017-11-28 08:08:37.457164 mon.0 [INF] pgmap v57410: 3736 pgs: 1750 active+undersized+degraded, 42 activating+undersized+degraded, 64 peering, 198 stale+active+clean, 1682 active+clean; 10343 bytes data, 4622 MB used, 214 TB / 214 TB avail; 141/791 objects degraded (17.826%)2017-11-28 08:08:38.466174 mon.0 [INF] pgmap v57411: 3736 pgs: 2450 active+undersized+degraded, 42 activating+undersized+degraded, 64 peering, 1180 active+clean; 10343 bytes data, 4643 MB used, 214 TB / 214 TB avail; 190/791 objects degraded (24.020%)2017-11-28 08:08:39.454811 mon.0 [INF] pgmap v57412: 3736 pgs: 2556 active+undersized+degraded, 1180 active+clean; 10343 bytes data, 4645 MB used, 214 TB / 214 TB avail; 193/791 objects degraded (24.399%)2017-11-28 08:08:45.202295 mon.0 [INF] HEALTH_WARN; 2556 pgs degraded; 2549 pgs stuck unclean; 2556 pgs undersized; recovery 193/791 objects degraded (24.399%); 11/59 in osds are down.... AFTER 5 MINUTES PGs REMAPPING HAS STARTED2017-11-28 08:12:45.205422 mon.0 [INF] HEALTH_WARN; 2556 pgs degraded; 2556 pgs stuck unclean; 2556 pgs undersized; recovery 193/791 objects degraded (24.399%); 11/59 in osds are down2017-11-28 08:12:51.570936 mon.0 [INF] pgmap v57446: 3736 pgs: 2556 active+undersized+degraded, 1180 active+clean; 10343 bytes data, 4632 MB used, 214 TB / 214 TB avail; 193/791 objects degraded (24.399%)2017-11-28 08:13:35.060583 mon.0 [INF] osd.49 out (down for 301.868797)2017-11-28 08:13:35.060723 mon.0 [INF] osd.50 out (down for 301.868797)2017-11-28 08:13:35.060753 mon.0 [INF] osd.51 out (down for 301.868797)2017-11-28 08:13:35.060783 mon.0 [INF] osd.52 out (down for 301.868796)2017-11-28 08:13:35.060812 mon.0 [INF] osd.53 out (down for 301.868796)2017-11-28 08:13:35.060842 mon.0 [INF] osd.54 out (down for 301.868796)2017-11-28 08:13:35.060870 mon.0 [INF] osd.55 out (down for 301.868795)2017-11-28 08:13:35.060928 mon.0 [INF] osd.56 out (down for 301.868795)2017-11-28 08:13:35.060958 mon.0 [INF] osd.57 out (down for 301.868795)2017-11-28 08:13:35.060990 mon.0 [INF] osd.58 out (down for 301.868795)2017-11-28 08:13:35.061021 mon.0 [INF] osd.59 out (down for 301.868794)2017-11-28 08:13:35.274737 mon.0 [INF] osdmap e541: 60 osds: 48 up, 48 in2017-11-28 08:13:35.276185 mon.0 [INF] pgmap v57447: 3736 pgs: 2556 active+undersized+degraded, 1180 active+clean; 10343 bytes data, 3773 MB used, 174 TB / 174 TB avail; 193/791 objects degraded (24.399%)2017-11-28 08:13:36.330316 mon.0 [INF] osdmap e542: 60 osds: 48 up, 48 in2017-11-28 08:13:36.334183 mon.0 [INF] pgmap v57448: 3736 pgs: 135 remapped+peering, 2421 active+undersized+degraded, 1180 active+clean; 10343 bytes data, 3775 MB used, 174 TB / 174 TB avail; 174/791 objects degraded (21.997%)2017-11-28 08:13:37.289319 mon.0 [INF] osdmap e543: 60 osds: 48 up, 48 in2017-11-28 08:13:37.326379 mon.0 [INF] pgmap v57449: 3736 pgs: 4 active+undersized+remapped, 900 peering, 329 remapped+peering, 1323 active+undersized+degraded, 1180 active+clean; 10343 bytes data, 3784 MB used, 174 TB / 174 TB avail; 69/791 objects degraded (8.723%); 65 B/s, 117 objects/s recovering2017-11-28 08:13:36.172666 osd.28 [INF] 8.6 starting backfill to osd.22 from (0'0,0'0] MAX to 538'465582017-11-28 08:13:36.174172 osd.28 [INF] 8.6 starting backfill to osd.47 from (0'0,0'0] MAX to 538'465582017-11-28 08:13:36.184611 osd.19 [INF] 8.5 starting backfill to osd.40 from (0'0,0'0] MAX to 538'529022017-11-28 08:13:36.190060 osd.24 [INF] 7.3 starting backfill to osd.8 from (0'0,0'0] MAX to 538'31722017-11-28 08:13:36.193337 osd.24 [INF] 7.3 starting backfill to osd.41 from (0'0,0'0] MAX to 538'31722017-11-28 08:13:37.517955 osd.21 [INF] 5.144 scrub starts2017-11-28 08:13:37.518701 osd.21 [INF] 5.144 scrub ok2017-11-28 08:13:38.235143 mon.0 [INF] osdmap e544: 60 osds: 48 up, 48 in2017-11-28 08:13:38.250128 mon.0 [INF] pgmap v57450: 3736 pgs: 37 activating, 7 activating+remapped, 4 active+undersized+remapped, 37 active, 902 peering, 87 active+remapped, 313 remapped+peering, 793 active+undersized+degraded, 1556 active+clean; 10343 bytes data, 3789 MB used, 174 TB / 174 TB avail; 40/791 objects degraded (5.057%); 5/791 objects misplaced (0.632%); 629 B/s, 167 objects/s recovering2017-11-28 08:13:36.157779 osd.18 [INF] 8.3 starting backfill to osd.17 from (0'0,0'0] MAX to 538'341582017-11-28 08:13:38.147555 osd.18 [INF] 5.203 deep-scrub starts2017-11-28 08:13:38.148310 osd.18 [INF] 5.203 deep-scrub ok2017-11-28 08:13:38.523380 osd.22 [INF] 5.235 scrub starts2017-11-28 08:13:38.524181 osd.22 [INF] 5.235 scrub ok2017-11-28 08:13:39.251064 mon.0 [INF] pgmap v57451: 3736 pgs: 37 activating, 7 activating+remapped, 4 active+undersized+remapped, 50 active, 903 peering, 117 active+remapped, 312 remapped+peering, 625 active+undersized+degraded, 1681 active+clean; 10343 bytes data, 3799 MB used, 174 TB / 174 TB avail; 25/791 objects degraded (3.161%); 5/791 objects misplaced (0.632%); 620 B/s, 0 keys/s, 58 objects/s recovering2017-11-28 08:13:36.110274 osd.4 [INF] 8.0 starting backfill to osd.14 from (0'0,0'0] MAX to 538'494822017-11-28 08:13:36.112128 osd.4 [INF] 8.0 starting backfill to osd.23 from (0'0,0'0] MAX to 538'494822017-11-28 08:13:36.127248 osd.4 [INF] 8.0 starting backfill to osd.37 from (0'0,0'0] MAX to 538'494822017-11-28 08:13:40.250559 mon.0 [INF] pgmap v57452: 3736 pgs: 37 activating, 7 activating+remapped, 4 active+undersized+remapped, 52 active, 903 peering, 123 active+remapped, 311 remapped+peering, 590 active+undersized+degraded, 1709 active+clean; 10343 bytes data, 3803 MB used, 174 TB / 174 TB avail; 25/791 objects degraded (3.161%); 5/791 objects misplaced (0.632%); 77 B/s, 0 keys/s, 14 objects/s recovering2017-11-28 08:13:36.153569 osd.2 [INF] 8.2 starting backfill to osd.6 from (0'0,0'0] MAX to 538'496462017-11-28 08:13:36.164089 osd.2 [INF] 8.2 starting backfill to osd.34 from (0'0,0'0] MAX to 538'496462017-11-28 08:13:36.217509 osd.10 [INF] 8.1 starting backfill to osd.0 from (0'0,0'0] MAX to 538'559462017-11-28 08:13:36.219512 osd.10 [INF] 8.1 starting backfill to osd.23 from (0'0,0'0] MAX to 538'559462017-11-28 08:13:37.806811 osd.10 [INF] 5.318 scrub starts2017-11-28 08:13:37.807563 osd.10 [INF] 5.318 scrub ok2017-11-28 08:13:36.235023 osd.45 [INF] 8.4 starting backfill to osd.2 from (0'0,0'0] MAX to 538'650042017-11-28 08:13:36.236576 osd.45 [INF] 8.4 starting backfill to osd.8 from (0'0,0'0] MAX to 538'650042017-11-28 08:13:39.607783 osd.3 [INF] 5.185 scrub starts2017-11-28 08:13:39.608687 osd.3 [INF] 5.185 scrub ok2017-11-28 08:13:41.357592 mon.0 [INF] pgmap v57453: 3736 pgs: 37 activating, 7 activating+remapped, 4 active+undersized+remapped, 75 active, 869 peering, 157 active+remapped, 174 remapped+peering, 540 active+undersized+degraded, 1873 active+clean; 10343 bytes data, 3813 MB used, 174 TB / 174 TB avail; 22/791 objects degraded (2.781%); 5/791 objects misplaced (0.632%); 87 B/s, 23 objects/s recovering2017-11-28 08:13:42.397617 mon.0 [INF] pgmap v57454: 3736 pgs: 146 active, 3 peering, 338 active+remapped, 540 active+undersized+degraded, 2709 active+clean; 10343 bytes data, 3835 MB used, 174 TB / 174 TB avail; 31/791 objects degraded (3.919%); 14/791 objects misplaced (1.770%); 2765 B/s, 27 keys/s, 56 objects/s recovering2017-11-28 08:13:37.396991 osd.14 [INF] 5.332 scrub starts2017-11-28 08:13:37.397496 osd.14 [INF] 5.332 scrub ok2017-11-28 08:13:42.524505 osd.6 [INF] 3.185 scrub starts2017-11-28 08:13:42.525389 osd.6 [INF] 3.185 scrub ok2017-11-28 08:13:43.385342 mon.0 [INF] pgmap v57455: 3736 pgs: 146 active, 338 active+remapped, 540 active+undersized+degraded, 2712 active+clean; 10343 bytes data, 3847 MB used, 174 TB / 174 TB avail; 31/791 objects degraded (3.919%); 14/791 objects misplaced (1.770%); 2768 B/s, 28 keys/s, 33 objects/s recovering2017-11-28 08:13:43.397979 osd.14 [INF] 8.0 scrub starts2017-11-28 08:13:43.401167 osd.14 [INF] 8.0 scrub ok2017-11-28 08:13:44.392089 mon.0 [INF] pgmap v57456: 3736 pgs: 146 active, 338 active+remapped, 540 active+undersized+degraded, 2712 active+clean; 10343 bytes data, 3848 MB used, 174 TB / 174 TB avail; 31/791 objects degraded (3.919%); 14/791 objects misplaced (1.770%)2017-11-28 08:13:45.206293 mon.0 [INF] HEALTH_WARN; 540 pgs degraded; 540 pgs stuck degraded; 1024 pgs stuck unclean; 540 pgs stuck undersized; 540 pgs undersized; recovery 31/791 objects degraded (3.919%); recovery 14/791 objects misplaced (1.770%)...2017-11-28 08:14:10.362591 osd.44 [WRN] 1 slow requests, 1 included below; oldest blocked for > 30.779132 secs2017-11-28 08:14:10.362600 osd.44 [WRN] slow request 30.779132 seconds old, received at 2017-11-28 08:13:39.583415: osd_op(client.4740.0:153303 4.31099063 (undecoded) ondisk+write+known_if_redirected e541) currently no flag points reached2017-11-28 08:14:11.579659 mon.0 [INF] pgmap v57474: 3736 pgs: 146 active, 338 active+remapped, 540 active+undersized+degraded, 2712 active+clean; 10343 bytes data, 3852 MB used, 174 TB / 174 TB avail; 31/791 objects degraded (3.919%); 14/791 objects misplaced (1.770%)2017-11-28 08:14:40.365929 osd.44 [WRN] 1 slow requests, 1 included below; oldest blocked for > 60.782471 secs2017-11-28 08:14:40.365934 osd.44 [WRN] slow request 60.782471 seconds old, received at 2017-11-28 08:13:39.583415: osd_op(client.4740.0:153303 4.31099063 (undecoded) ondisk+write+known_if_redirected e541) currently no flag points reached2017-11-28 08:14:45.207183 mon.0 [INF] HEALTH_WARN; 540 pgs degraded; 540 pgs stuck degraded; 1024 pgs stuck unclean; 540 pgs stuck undersized; 540 pgs undersized; 1 requests are blocked > 32 sec; recovery 31/791 objects degraded (3.919%); recovery 14/791 objects misplaced (1.770%)2017-11-28 08:14:46.657287 mon.0 [INF] pgmap v57478: 3736 pgs: 146 active, 338 active+remapped, 540 active+undersized+degraded, 2712 active+clean; 10343 bytes data, 3852 MB used, 174 TB / 174 TB avail; 31/791 objects degraded (3.919%); 14/791 objects misplaced (1.770%)2017-11-28 08:15:40.372583 osd.44 [WRN] 1 slow requests, 1 included below; oldest blocked for > 120.789122 secs2017-11-28 08:15:40.372589 osd.44 [WRN] slow request 120.789122 seconds old, received at 2017-11-28 08:13:39.583415: osd_op(client.4740.0:153303 4.31099063 (undecoded) ondisk+write+known_if_redirected e541) currently no flag points reached2017-11-28 08:15:56.664417 mon.0 [INF] pgmap v57479: 3736 pgs: 146 active, 338 active+remapped, 540 active+undersized+degraded, 2712 active+clean; 10343 bytes data, 3852 MB used, 174 TB / 174 TB avail; 31/791 objects degraded (3.919%); 14/791 objects misplaced (1.770%)# NOW CEPH STATUS ISroot@host01:~# ceph statuscluster a6f73750-1972-47f6-bcf5-a99753be65adhealth HEALTH_WARN540 pgs degraded540 pgs stuck degraded1024 pgs stuck unclean540 pgs stuck undersized540 pgs undersized1 requests are blocked > 32 secrecovery 31/791 objects degraded (3.919%)recovery 14/791 objects misplaced (1.770%)monmap e2: 3 mons at {host01=10.212.32.23:6789/0,host02=10.212.32.24:6789/0,host03=10.212.32.25:6789/0}election epoch 22, quorum 0,1,2 host01,host02,host03osdmap e544: 60 osds: 48 up, 48 in; 1024 remapped pgsflags sortbitwise,require_jewel_osdspgmap v57508: 3736 pgs, 19 pools, 10343 bytes data, 241 objects3786 MB used, 174 TB / 174 TB avail31/791 objects degraded (3.919%)14/791 objects misplaced (1.770%)2712 active+clean540 active+undersized+degraded338 active+remapped146 activeroot@host01:~## LOOKS THAT 338 PGs IN ERASURE CODED POOLS HAVE BEEN REMAPPED# I DONT GET WHY 540 PGs STILL ENCOUNTER active+undersized+degraded STATEroot@host01:~# ceph pg dump pgs_brief |grep 'active+remapped'dumped pgs_brief in format plain....root@host01:~# egrep '16.6f|16.6e' PGs_on_HOST_host0516.6f active+clean [43,33,2,59,12] 43 [43,33,2,59,12] 4316.6e active+clean [10,5,49,35,41] 10 [10,5,49,35,41] 10root@host01:~#root@host01:~# ceph pg dump pgs_brief |grep 'active+undersized+degraded'dumped pgs_brief in format plain...root@host01:~# egrep '19.6c|17.6e' PGs_on_HOST_host0519.6c active+clean [24,20,19,58,46] 24 [24,20,19,58,46] 2417.6e active+clean [19,59,36,31,5] 19 [19,59,36,31,5] 19root@host01:~## POOLS DETAILSroot@host01:~# ceph osd lspools0 rbd,1 .rgw.root,2 vms,3 images,4 default.rgw.control,5 volumes,6 default.rgw.data.root.old,7 default.rgw.gc,8 default.rgw.log,9 default.rgw.users.uid,10 default.rgw.users.keys,11 default.rgw.users.email,12 default.rgw.buckets.index,13 default.rgw.usage,14 default.rgw.buckets.data.old,15 ecpool_3_2,16 default.rgw.data.root,17 default.rgw.data.root.new01,19 default.rgw.buckets.data,rbd size: 3 pgp_num: 64.rgw.root size: 3 pgp_num: 8vms size: 3 pgp_num: 1024images size: 3 pgp_num: 512default.rgw.control size: 3 pgp_num: 8volumes size: 3 pgp_num: 1024default.rgw.data.root.old size: 3 pgp_num: 8default.rgw.gc size: 3 pgp_num: 8default.rgw.log size: 3 pgp_num: 8default.rgw.users.uid size: 3 pgp_num: 8default.rgw.users.keys size: 3 pgp_num: 8default.rgw.users.email size: 3 pgp_num: 8default.rgw.buckets.index size: 3 pgp_num: 8default.rgw.usage size: 3 pgp_num: 8default.rgw.buckets.data.old size: 3 pgp_num: 8ecpool_3_2 size: 5 pgp_num: 256default.rgw.data.root size: 5 pgp_num: 256default.rgw.data.root.new01 size: 5 pgp_num: 256default.rgw.buckets.data size: 5 pgp_num: 256# EC pools use below profileroot@host01:~# ceph osd erasure-code-profile get ec_profile_k_3_m_2jerasure-per-chunk-alignment=falsek=3m=2plugin=jerasureruleset-failure-domain=hostruleset-root=defaulttechnique=reed_sol_vanw=8root@host01:~## PGs that are in active+remapped or active+undersized+degraded state belong to erasure coded pools onlyroot@host01:~# ceph pg dump pgs_brief |grep 'active+remapped' |cut -d '.' -f1 |sort |uniqdumped pgs_brief in format plain15161719root@host01:~# ceph pg dump pgs_brief |grep 'active+undersized+degraded' |cut -d '.' -f1 |sort |uniqdumped pgs_brief in format plain15161719# FINALLY, CRUSH MAP ISroot@host01:~# cat crushmap.txt# begin crush maptunable choose_local_tries 0tunable choose_local_fallback_tries 0tunable choose_total_tries 50tunable chooseleaf_descend_once 1tunable chooseleaf_vary_r 1tunable straw_calc_version 1# devicesdevice 0 osd.0device 1 osd.1device 2 osd.2device 3 osd.3device 4 osd.4device 5 osd.5device 6 osd.6device 7 osd.7device 8 osd.8device 9 osd.9device 10 osd.10device 11 osd.11device 12 osd.12device 13 osd.13device 14 osd.14device 15 osd.15device 16 osd.16device 17 osd.17device 18 osd.18device 19 osd.19device 20 osd.20device 21 osd.21device 22 osd.22device 23 osd.23device 24 osd.24device 25 osd.25device 26 osd.26device 27 osd.27device 28 osd.28device 29 osd.29device 30 osd.30device 31 osd.31device 32 osd.32device 33 osd.33device 34 osd.34device 35 osd.35device 36 osd.36device 37 osd.37device 38 osd.38device 39 osd.39device 40 osd.40device 41 osd.41device 42 osd.42device 43 osd.43device 44 osd.44device 45 osd.45device 46 osd.46device 47 osd.47device 48 osd.48device 49 osd.49device 50 osd.50device 51 osd.51device 52 osd.52device 53 osd.53device 54 osd.54device 55 osd.55device 56 osd.56device 57 osd.57device 58 osd.58device 59 osd.59# typestype 0 osdtype 1 hosttype 2 chassistype 3 racktype 4 rowtype 5 pdutype 6 podtype 7 roomtype 8 datacentertype 9 regiontype 10 root# bucketshost host01 {id -2 # do not change unnecessarily# weight 43.637alg strawhash 0 # rjenkins1item osd.0 weight 3.636item osd.3 weight 3.636item osd.4 weight 3.636item osd.6 weight 3.636item osd.8 weight 3.636item osd.10 weight 3.636item osd.12 weight 3.636item osd.14 weight 3.636item osd.16 weight 3.636item osd.19 weight 3.636item osd.22 weight 3.636item osd.25 weight 3.636}rack Rack01 {id -3 # do not change unnecessarily# weight 43.637alg strawhash 0 # rjenkins1item host01 weight 43.637}host host02 {id -7 # do not change unnecessarily# weight 43.637alg strawhash 0 # rjenkins1item osd.1 weight 3.636item osd.2 weight 3.636item osd.5 weight 3.636item osd.7 weight 3.636item osd.9 weight 3.636item osd.11 weight 3.636item osd.13 weight 3.636item osd.15 weight 3.636item osd.17 weight 3.636item osd.20 weight 3.636item osd.23 weight 3.636item osd.26 weight 3.636}rack Rack02 {id -8 # do not change unnecessarily# weight 43.637alg strawhash 0 # rjenkins1item host02 weight 43.637}host host03 {id -9 # do not change unnecessarily# weight 43.637alg strawhash 0 # rjenkins1item osd.18 weight 3.636item osd.21 weight 3.636item osd.24 weight 3.636item osd.27 weight 3.636item osd.28 weight 3.636item osd.29 weight 3.636item osd.30 weight 3.636item osd.31 weight 3.636item osd.32 weight 3.636item osd.33 weight 3.636item osd.34 weight 3.636item osd.35 weight 3.636}host host04 {id -11 # do not change unnecessarily# weight 43.637alg strawhash 0 # rjenkins1item osd.36 weight 3.636item osd.37 weight 3.636item osd.38 weight 3.636item osd.39 weight 3.636item osd.40 weight 3.636item osd.41 weight 3.636item osd.42 weight 3.636item osd.43 weight 3.636item osd.44 weight 3.636item osd.45 weight 3.636item osd.46 weight 3.636item osd.47 weight 3.636}host host05 {id -12 # do not change unnecessarily# weight 43.637alg strawhash 0 # rjenkins1item osd.48 weight 3.636item osd.49 weight 3.636item osd.50 weight 3.636item osd.51 weight 3.636item osd.52 weight 3.636item osd.53 weight 3.636item osd.54 weight 3.636item osd.55 weight 3.636item osd.56 weight 3.636item osd.57 weight 3.636item osd.58 weight 3.636item osd.59 weight 3.636}rack Rack03 {id -10 # do not change unnecessarily# weight 130.910alg strawhash 0 # rjenkins1item host03 weight 43.637item host04 weight 43.637item host05 weight 43.637}room MyRoom {id -4 # do not change unnecessarily# weight 218.184alg strawhash 0 # rjenkins1item Rack01 weight 43.637item Rack02 weight 43.637item Rack03 weight 130.910}datacenter MyDC {id -5 # do not change unnecessarily# weight 218.184alg strawhash 0 # rjenkins1item MyRoom weight 218.184}region MyRegion {id -6 # do not change unnecessarily# weight 218.184alg strawhash 0 # rjenkins1item MyDC weight 218.184}root default {id -1 # do not change unnecessarily# weight 218.184alg strawhash 0 # rjenkins1item MyRegion weight 218.184}# rulesrule replicated_ruleset {ruleset 0type replicatedmin_size 1max_size 10step take defaultstep chooseleaf firstn 0 type hoststep emit}rule ecpool_3_2 {ruleset 1type erasuremin_size 3max_size 5step set_chooseleaf_tries 5step set_choose_tries 100step take defaultstep chooseleaf indep 0 type hoststep emit}rule default.rgw.data.root.new {ruleset 2type erasuremin_size 3max_size 5step set_chooseleaf_tries 5step set_choose_tries 100step take defaultstep chooseleaf indep 0 type hoststep emit}rule default.rgw.data.root.new01 {ruleset 3type erasuremin_size 3max_size 5step set_chooseleaf_tries 5step set_choose_tries 100step take defaultstep chooseleaf indep 0 type hoststep emit}rule default.rgw.buckets.data.new {ruleset 4type erasuremin_size 3max_size 5step set_chooseleaf_tries 5step set_choose_tries 100step take defaultstep chooseleaf indep 0 type hoststep emit}# end crush maproot@host01:~#Jakub
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com