On Fri, 31 Mar 2017, nokia ceph wrote: > Hello Brad, > Many thanks of the info :) > > ENV:-- Kracken - bluestore - EC 4+1 - 5 node cluster : RHEL7 > > What is the status of the down+out osd? Only one osd osd.6 down and out from > cluster. > What role did/does it play? Mostimportantly, is it osd.6? Yes, due to > underlying I/O error issue we removed this device from the cluster. Is the device completely destroyed or is it only returning errors when reading certain data? It is likely that some (or all) of the incomplete PGs can be extracted from the drive if the bad sector(s) don't happen to affect those pgs. The ceph-objectstore-tool --op export command can be used for this (extract it from the affected drive and add it to some other osd). > I put this parameter " osd_find_best_info_ignore_history_les = true" in > ceph.conf, and find those 22 PG's were changed to "down+remapped" . Now all > are reverted to "remapped+incomplete" state. This is usually not a great idea unless you're out of options, by the way! > #ceph pg stat 2> /dev/null > v2731828: 4096 pgs: 1 incomplete, 21 remapped+incomplete, 4074 active+clean; > 268 TB data, 371 TB used, 267 TB / 638 TB avail > > ## ceph -s > 2017-03-30 19:02:14.350242 7f8b0415f700 -1 WARNING: the following dangerous > and experimental features are enabled: bluestore,rocksdb > 2017-03-30 19:02:14.366545 7f8b0415f700 -1 WARNING: the following dangerous > and experimental features are enabled: bluestore,rocksdb > cluster bd8adcd0-c36d-4367-9efe-f48f5ab5f108 > health HEALTH_ERR > 22 pgs are stuck inactive for more than 300 seconds > 22 pgs incomplete > 22 pgs stuck inactive > 22 pgs stuck unclean > monmap e2: 5 mons at{au-adelaide=10.50.21.24:6789/0,au-brisbane=10.50.21.22:6789/0,au-canberra= > 10.50.21.23:6789/0,au-melbourne=10.50.21.21:6789/0,au-sydney=10.50.21.20:67 > 89/0} > election epoch 180, quorum 0,1,2,3,4 > au-sydney,au-melbourne,au-brisbane,au-canberra,au-adelaide > mgr active: au-adelaide > osdmap e6506: 117 osds: 117 up, 117 in; 21 remapped pgs > flags sortbitwise,require_jewel_osds,require_kraken_osds > pgmap v2731828: 4096 pgs, 1 pools, 268 TB data, 197 Mobjects > 371 TB used, 267 TB / 638 TB avail > 4074 active+clean > 21 remapped+incomplete > 1 incomplete > > > ## ceph osd dump 2>/dev/null | grep cdvr > pool 1 'cdvr_ec' erasure size 5 min_size 4 crush_ruleset 1 object_hash > rjenkins pg_num 4096 pgp_num 4096 last_change 456 flags > hashpspool,nodeep-scrub stripe_width 65536 > > Inspecting affected PG 1.e4b > > # ceph pg dump 2> /dev/null | grep 1.e4b > 1.e4b 50832 0 0 0 0 73013340821 > 10006 10006 remapped+incomplete 2017-03-30 14:14:26.297098 3844'161662 > 6506:325748 [113,66,15,73,103] 113 [NONE,NONE,NONE,73,NONE] > 73 1643'139486 2017-03-21 04:56:16.683953 0'0 2017-02-21 > 10:33:50.012922 > > When I trigger below command. > > #ceph pg force_create_pg 1.e4b > pg 1.e4b now creating, ok > > As it went to creating state, no change after that. Can you explain why this > PG showing null values after triggering "force_create_pg",? > > ]# ceph pg dump 2> /dev/null | grep 1.e4b > 1.e4b 0 0 0 0 0 0 > 0 0 creating 2017-03-30 19:07:00.982178 0'0 > 0:0 [] -1 [] > -1 0'0 0.000000 0'0 > 0.000000 CRUSH isn't mapping the PG to any OSDs, so there is nowhere to create it, it seems? What does 'ceph pg map <pgid>' show? > Then I triggered below command > > # ceph pg repair 1.e4b > Error EAGAIN: pg 1.e4b has no primary osd --<< > > Could you please provide answer for below queries. > > 1. How to fix this "incomplete+remapped" PG issue, here all OSD's were up > and running and affected OSD marked out and removed from the cluster. To recover the data, you need to find surviving shards of the PG. ceph-objectstore-tool on the "failed" disk is one option, but since this is a 4+2 code there should have been another copy that got lost along the line... do you know where it is? > 2. Will reduce min_size helps? currently it set to 4. Could you please > explain what is the impact if we reduce min_size for the current config EC > 4+1 You can't reduce it below 4 since it's a 4+2 code. By default we set it as 5 (k+1) so that you won't write new data to the PG if a single additional failure could lead you to lose those writes. > 3. Is there any procedure to safely remove an affected PG? As per my > understanding I'm aware about this command. > > === > #ceph-objectstore-tool --data-path /var/lib/ceph/osd/ceph --pgid 1.e4b --op > remove > === > > Awaiting for your suggestions to proceed. If you don't need the data and just want to recreate the pg empty, then the procedure is to remove any surviving fragments and then do force_create_pg. It looks like you need to figure out why the pgid isn't mapping to any OSDs first, though. sage > > Thanks > > > > > > > On Thu, Mar 30, 2017 at 7:32 AM, Brad Hubbard <bhubbard@xxxxxxxxxx> wrote: > > > On Thu, Mar 30, 2017 at 4:53 AM, nokia ceph > <nokiacephusers@xxxxxxxxx> wrote: > > Hello, > > > > Env:- > > 5 node, EC 4+1 bluestore kraken v11.2.0 , RHEL7.2 > > > > As part of our resillency testing with kraken bluestore, we > face more PG's > > were in incomplete+remapped state. We tried to repair each PG > using "ceph pg > > repair <pgid>" still no luck. Then we planned to remove > incomplete PG's > > using below procedure. > > > > > > #ceph health detail | grep 1.e4b > > pg 1.e4b is remapped+incomplete, acting > [2147483647,66,15,73,2147483647] > > (reducing pool cdvr_ec min_size from 4 may help; search > ceph.com/docs for > > 'incomplete') > > "Incomplete Ceph detects that a placement group is missing > information about > writes that may have occurred, or does not have any healthy > copies. If you see > this state, try to start any failed OSDs that may contain the > needed > information." > > > > > Here we shutdown the OSD's 66,15 and 73 then proceeded with > below operation. > > > > #ceph-objectstore-tool --data-path /var/lib/ceph/osd/ceph-135 > --op list-pgs > > #ceph-objectstore-tool --data-path /var/lib/ceph/osd/ceph-135 > --pgid 1.e4b > > --op remove > > > > Please confirm that we are following the correct procedure to > removal of > > PG's > > There are multiple threads about that on this very list "pgs > stuck inactive" > recently for example. > > > > > #ceph pg stat > > v2724830: 4096 pgs: 1 active+clean+scrubbing+deep+repair, 1 > down+remapped, > > 21 remapped+incomplete, 4073 active+clean; 268 TB data, 371 TB > used, 267 TB > > / 638 TB avail > > > > # ceph -s > > 2017-03-29 18:23:44.288508 7f8c2b8e5700 -1 WARNING: the > following dangerous > > and experimental features are enabled: bluestore,rocksdb > > 2017-03-29 18:23:44.304692 7f8c2b8e5700 -1 WARNING: the > following dangerous > > and experimental features are enabled: bluestore,rocksdb > > cluster bd8adcd0-c36d-4367-9efe-f48f5ab5f108 > > health HEALTH_ERR > > 22 pgs are stuck inactive for more than 300 > seconds > > 1 pgs down > > 21 pgs incomplete > > 1 pgs repair > > 22 pgs stuck inactive > > 22 pgs stuck unclean > > monmap e2: 5 mons at > >{au-adelaide=10.50.21.24:6789/0,au-brisbane=10.50.21.22:6789/0,au-canberra= > 10.50.21.23:6789/0,au-melbourne=10.50.21.21:6789/0,au-sydney=10.50.21.20:67 > 89/0} > > election epoch 172, quorum 0,1,2,3,4 > > au-sydney,au-melbourne,au-brisbane,au-canberra,au-adelaide > > mgr active: au-brisbane > > osdmap e6284: 118 osds: 117 up, 117 in; 22 remapped pgs > > What is the status of the down+out osd? What role did/does it > play? Most > importantly, is it osd.6? > > > flags > sortbitwise,require_jewel_osds,require_kraken_osds > > pgmap v2724830: 4096 pgs, 1 pools, 268 TB data, 197 > Mobjects > > 371 TB used, 267 TB / 638 TB avail > > 4073 active+clean > > 21 remapped+incomplete > > 1 down+remapped > > 1 active+clean+scrubbing+deep+repair > > > > > > #ceph osd dump | grep pool > > pool 1 'cdvr_ec' erasure size 5 min_size 4 crush_ruleset 1 > object_hash > > rjenkins pg_num 4096 pgp_num 4096 last_change 456 flags > > hashpspool,nodeep-scrub stripe_width 65536 > > > > > > > > Can you please suggest is there any way to wipe out these > incomplete PG's. > > See the thread previously mentioned. Take note of the > force_create_pg step. > > > Why ceph pg repair failed in this scenerio? > > How to recover incomplete PG's to active state. > > > > pg query for the affected PG ended with this error. Can you > please explain > > what is meant by this ? > > --- > > "15(2)", > > "66(1)", > > "73(3)", > > "103(4)", > > "113(0)" > > ], > > "down_osds_we_would_probe": [ > > 6 > > ], > > "peering_blocked_by": [], > > "peering_blocked_by_detail": [ > > { > > "detail": > "peering_blocked_by_history_les_bound" > > } > > ---- > > During multiple intervals osd 6 was in the up/acting set, for > example; > > { > "first": 1608, > "last": 1645, > "maybe_went_rw": 1, > "up": [ > 113, > 6, > 15, > 73, > 103 > ], > "acting": [ > 113, > 6, > 15, > 73, > 103 > ], > "primary": 113, > "up_primary": 113 > }, > > Because we may have gone rw during that interval we need to > query it and it is blocking progress. > > "blocked_by": [ > 6 > ], > > Setting osd_find_best_info_ignore_history_les to true may help > but then you may > need to mark the missing OSD lost or perform some other trickery > (and this . I > suspect your min_size is too low, especially for a cluster of > this size, but EC > is not an area I know extensively so I can't say definitively. > Some of your > questions may be better suited to the ceph-devel mailing list by > the way. > > > > > Attaching "ceph pg 1.e4b query > /tmp/1.e4b-pg.txt" file with > this mail. > > > > Thanks > > > > _______________________________________________ > > ceph-users mailing list > > ceph-users@xxxxxxxxxxxxxx > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > > > -- > Cheers, > Brad > > > >
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com