Re: PG down, due to 3 OSD failing

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Ciao Dan,
this is what I did with chunk s3, copying it from osd.121 to osd.176 (which is managed by the same host).

But still
pg 85.25 is stuck stale for 85029.707069, current state stale+down+remapped, last acting [2147483647,2147483647,96,2147483647,2147483647]

So "health detail" apparently plainly ignores osd.176: moreover, its output only shows OSD 96, but I just checked again and the other chunks are still on OSDs 56,64,140,159 which are all "up".

By the way, you talk about a "bug" in your message: do you have any specific one in mind, or was it just a generic synonym for "problem"?
By the way, I uploaded here:
	https://pastebin.ubuntu.com/p/dTfPkMb7mD/
a few hundreds of lines from one of the failed OSDs upon "activate --all".

  Thanks

			Fulvio

On 29/03/2022 10:53, Dan van der Ster wrote:
Hi Fulvio,

I don't think upmap will help -- that is used to remap where data
should be "up", but your problem is more that the PG chunks are not
going active due to the bug.

What happens if you export one of the PG chunks then import it to
another OSD -- does that chunk become active?

-- dan



On Tue, Mar 29, 2022 at 10:51 AM Fulvio Galeazzi
<fulvio.galeazzi@xxxxxxx> wrote:

Hallo again Dan, I am afraid I'd need a little more help, please...

Current status is as follows.

This is where I moved the chunk which was on osd.121:
~]# ceph-objectstore-tool --data-path /var/lib/ceph/osd/cephpa1-176
--no-mon-config --op list-pgs  | grep ^85\.25
85.25s3

while other chunks are (server - osd.id):
=== r2srv101.pa1.box.garr - 96
85.25s2
=== r2srv100.pa1.box.garr - 121  <-- down, chunk is on osd.176
85.25s3
=== r2srv100.pa1.box.garr - 159
85.25s1
85.25s4
=== r1-sto09.pa1.box.garr - 56
85.25s4
=== r1-sto09.pa1.box.garr - 64
85.25s0
=== r3srv15.pa1.box.garr - 140
85.25s1

Health detail shows that just one chunk can be found (if I understand
the output correctly):

~]# ceph health detail | grep 85\.25
      pg 85.25 is stuck stale for 5680.315732, current state
stale+down+remapped, last acting
[2147483647,2147483647,96,2147483647,2147483647]

Can I run some magic upmap command to explain my cluster where all the
chunks are? What would be the right syntax?
Little additional problem: I see s1 and s4 twice... I guess this was due
to remapping, as I was adding disks to the cluster: which one is the
right copy?

    Thanks!

                         Fulvio

Il 3/29/2022 9:35 AM, Fulvio Galeazzi ha scritto:
Thanks a lot, Dan!

  > The EC pgs have a naming convention like 85.25s1 etc.. for the various
  > k/m EC shards.

That was the bit of information I was missing... I was looking for the
wrong object.
I can now go on and export/import that one PGid chunk.

    Thanks again!

              Fulvio

On 28/03/2022 16:27, Dan van der Ster wrote:
Hi Fulvio,

You can check (offline) which PGs are on an OSD with the list-pgs op,
e.g.

ceph-objectstore-tool  --data-path /var/lib/ceph/osd/cephpa1-158/
--op list-pgs


-- dan


On Mon, Mar 28, 2022 at 2:29 PM Fulvio Galeazzi
<fulvio.galeazzi@xxxxxxx> wrote:

Hallo,
       all of a sudden, 3 of my OSDs failed, showing similar messages in
the log:

.....
       -5> 2022-03-28 14:19:02.451 7fc20fe99700  5 osd.145 pg_epoch:
616454 pg[70.2c6s1( empty local-lis/les=612106/612107 n=0
ec=148456/148456 lis/c
    612106/612106 les/c/f 612107/612107/0 612106/612106/612101)
[168,145,102,96,112,124,128,134,56,34]p168(0) r=1 lpr=616429 crt=0'0
unknown mbc={}]
enter Started
       -4> 2022-03-28 14:19:02.451 7fc20fe99700  5 osd.145 pg_epoch:
616454 pg[70.2c6s1( empty local-lis/les=612106/612107 n=0
ec=148456/148456 lis/c
    612106/612106 les/c/f 612107/612107/0 612106/612106/612101)
[168,145,102,96,112,124,128,134,56,34]p168(0) r=1 lpr=616429 crt=0'0
unknown mbc={}]
enter Start
       -3> 2022-03-28 14:19:02.451 7fc20fe99700  1 osd.145 pg_epoch:
616454 pg[70.2c6s1( empty local-lis/les=612106/612107 n=0
ec=148456/148456 lis/c
    612106/612106 les/c/f 612107/612107/0 612106/612106/612101)
[168,145,102,96,112,124,128,134,56,34]p168(0) r=1 lpr=616429 crt=0'0
unknown mbc={}]
state<Start>: transitioning to Stray
       -2> 2022-03-28 14:19:02.451 7fc20fe99700  5 osd.145 pg_epoch:
616454 pg[70.2c6s1( empty local-lis/les=612106/612107 n=0
ec=148456/148456 lis/c
    612106/612106 les/c/f 612107/612107/0 612106/612106/612101)
[168,145,102,96,112,124,128,134,56,34]p168(0) r=1 lpr=616429 crt=0'0
unknown mbc={}]
exit Start 0.000008 0 0.000000
       -1> 2022-03-28 14:19:02.451 7fc20fe99700  5 osd.145 pg_epoch:
616454 pg[70.2c6s1( empty local-lis/les=612106/612107 n=0
ec=148456/148456 lis/c
    612106/612106 les/c/f 612107/612107/0 612106/612106/612101)
[168,145,102,96,112,124,128,134,56,34]p168(0) r=1 lpr=616429 crt=0'0
unknown mbc={}]
enter Started/Stray
        0> 2022-03-28 14:19:02.451 7fc20f698700 -1 *** Caught signal
(Aborted) **
    in thread 7fc20f698700 thread_name:tp_osd_tp

    ceph version 14.2.22 (ca74598065096e6fcbd8433c8779a2be0c889351)
nautilus (stable)
    1: (()+0x12ce0) [0x7fc2327dcce0]
    2: (gsignal()+0x10f) [0x7fc231452a4f]
    3: (abort()+0x127) [0x7fc231425db5]
    4: (ceph::__ceph_abort(char const*, int, char const*,
std::__cxx11::basic_string<char, std::char_traits<char>,
std::allocator<char> > const&)+0x1b4) [0x55b8139cb671]
    5: (PG::check_past_interval_bounds() const+0xc16) [0x55b813b586f6]
    6: (PG::RecoveryState::Reset::react(PG::AdvMap const&)+0x3e8)
[0x55b813b963d8]
    7: (boost::statechart::simple_state<PG::RecoveryState::Reset,
PG::RecoveryState::RecoveryMachine, boost::mpl::list<mpl_::na, mpl_::na,
mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na,
mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na,
mpl_::na, mpl_::na, mpl_::na, mpl_::na>,
(boost::statechart::history_mode)0>::react_impl(boost::statechart::event_base

const&, void const*)+0x7d) [0x55b813bdd32d]
    8: (PG::handle_advance_map(std::shared_ptr<OSDMap const>,
std::shared_ptr<OSDMap const>, std::vector<int, std::allocator<int> >&,
int, std::vector<int, std::allocator<int> >&, int,
PG::RecoveryCtx*)+0x39d) [0x55b813b7b5fd]
    9: (OSD::advance_pg(unsigned int, PG*, ThreadPool::TPHandle&,
PG::RecoveryCtx*)+0x2e9) [0x55b813ad14e9]
    10: (OSD::dequeue_peering_evt(OSDShard*, PG*,
std::shared_ptr<PGPeeringEvent>, ThreadPool::TPHandle&)+0xaa)
[0x55b813ae345a]
    11: (PGPeeringItem::run(OSD*, OSDShard*, boost::intrusive_ptr<PG>&,
ThreadPool::TPHandle&)+0x55) [0x55b813d66c15]
    12: (OSD::ShardedOpWQ::_process(unsigned int,
ceph::heartbeat_handle_d*)+0x1366) [0x55b813adff46]
    13: (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x5c4)
[0x55b8140dc944]
    14: (ShardedThreadPool::WorkThreadSharded::entry()+0x14)
[0x55b8140df514]
    15: (()+0x81cf) [0x7fc2327d21cf]
    16: (clone()+0x43) [0x7fc23143dd83]

Trying to "activate --all", rebotting server, and such, did not help.

I am now stuck with one PG (85.25) down, find below the output from
"query".

The PG belongs to a 3+2 erasure-coded pool.
As the devices corresponding to the 3 down OSDs are properly mounted, is
there a way to get PG.ID=85.25 from the devices and copy it elsewhere?
Actually, I tried to find 85.25 in the 3 down OSDs with command:
~]# ceph-objectstore-tool  --data-path /var/lib/ceph/osd/cephpa1-158/
--no-mon-config --pgid 85.25 --op export --file /tmp/pg_85-25
PG '85.25' not found
     which puzzled me... is there a way to search such PG.ID over the
whole cluster?

     Thanks for your help!

                          Fulvio

====================

~]# ceph --cluster cephpa1 health detail | grep down
.....
PG_AVAILABILITY Reduced data availability: 1 pg inactive, 1 pg down
       pg 85.25 is down+remapped, acting
[2147483647,2147483647,96,2147483647,2147483647]
~]# ceph --cluster cephpa1 pg 85.25 query
{
       "state": "down+remapped",
       "snap_trimq": "[]",
       "snap_trimq_len": 0,
       "epoch": 617667,
       "up": [
           2147483647,
           2147483647,
           2147483647,
           2147483647,
           2147483647
       ],
       "acting": [
           2147483647,
           2147483647,
           96,
           2147483647,
           2147483647
       ],
       "info": {
           "pgid": "85.25s2",
           "last_update": "606021'521273",
           "last_complete": "606021'521273",
           "log_tail": "605873'518175",
           "last_user_version": 521273,
           "last_backfill": "MAX",
           "last_backfill_bitwise": 1,
           "purged_snaps": [],
           "history": {
               "epoch_created": 478130,
               "epoch_pool_created": 478130,
               "last_epoch_started": 611605,
               "last_interval_started": 611604,
               "last_epoch_clean": 595152,
               "last_interval_clean": 595151,
               "last_epoch_split": 0,
               "last_epoch_marked_full": 0,
               "same_up_since": 617661,
               "same_interval_since": 617661,
               "same_primary_since": 616519,
               "last_scrub": "597862'517979",
               "last_scrub_stamp": "2022-03-15 05:08:42.561720",
               "last_deep_scrub": "597454'511650",
               "last_deep_scrub_stamp": "2022-03-09 18:05:34.205121",
               "last_clean_scrub_stamp": "2022-03-15 05:08:42.561720"
           },
           "stats": {
               "version": "606021'521273",
               "reported_seq": "1733225",
               "reported_epoch": "617667",
               "state": "down+remapped",
               "last_fresh": "2022-03-28 14:17:48.920542",
               "last_change": "2022-03-28 14:15:42.025193",
               "last_active": "2022-03-25 15:24:42.557667",
               "last_peered": "2022-03-25 15:24:25.912029",
               "last_clean": "2022-03-15 01:50:02.979366",
               "last_became_active": "2022-03-25 12:44:52.379711",
               "last_became_peered": "2022-03-25 12:44:52.379711",
               "last_unstale": "2022-03-28 14:17:48.920542",
               "last_undegraded": "2022-03-28 14:17:48.920542",
               "last_fullsized": "2022-03-28 14:17:48.920542",
               "mapping_epoch": 617661,
               "log_start": "605873'518175",
               "ondisk_log_start": "605873'518175",
               "created": 478130,
               "last_epoch_clean": 595152,
               "parent": "0.0",
               "parent_split_bits": 0,
               "last_scrub": "597862'517979",
               "last_scrub_stamp": "2022-03-15 05:08:42.561720",
               "last_deep_scrub": "597454'511650",
               "last_deep_scrub_stamp": "2022-03-09 18:05:34.205121",
               "last_clean_scrub_stamp": "2022-03-15 05:08:42.561720",
               "log_size": 3098,
               "ondisk_log_size": 3098,
               "stats_invalid": false,
               "dirty_stats_invalid": false,
               "omap_stats_invalid": false,
               "hitset_stats_invalid": false,
               "hitset_bytes_stats_invalid": false,
               "pin_stats_invalid": false,
               "manifest_stats_invalid": false,
               "snaptrimq_len": 0,
               "stat_sum": {
                   "num_bytes": 165108592640,
                   "num_objects": 39402,
                   "num_object_clones": 0,
                   "num_object_copies": 197010,
                   "num_objects_missing_on_primary": 0,
                   "num_objects_missing": 0,
                   "num_objects_degraded": 0,
                   "num_objects_misplaced": 0,
                   "num_objects_unfound": 0,
                   "num_objects_dirty": 39402,
                   "num_whiteouts": 0,
                   "num_read": 338705,
                   "num_read_kb": 124297874,
                   "num_write": 316971,
                   "num_write_kb": 311968402,
                   "num_scrub_errors": 0,
                   "num_shallow_scrub_errors": 0,
                   "num_deep_scrub_errors": 0,
                   "num_objects_recovered": 196778,
                   "num_bytes_recovered": 824577241088,
                   "num_keys_recovered": 0,
                   "num_objects_omap": 0,
                   "num_objects_hit_set_archive": 0,
                   "num_bytes_hit_set_archive": 0,
                   "num_flush": 0,
                   "num_flush_kb": 0,
                   "num_evict": 0,
                   "num_evict_kb": 0,
                   "num_promote": 0,
                   "num_flush_mode_high": 0,
                   "num_flush_mode_low": 0,
                   "num_evict_mode_some": 0,
                   "num_evict_mode_full": 0,
                   "num_objects_pinned": 0,
                   "num_legacy_snapsets": 0,
                   "num_large_omap_objects": 0,
                   "num_objects_manifest": 0,
                   "num_omap_bytes": 0,
                   "num_omap_keys": 0,
                   "num_objects_repaired": 0
               },
               "up": [
                   2147483647,
                   2147483647,
                   2147483647,
                   2147483647,
                   2147483647
               ],
               "acting": [
                   2147483647,
                   2147483647,
                   96,
                   2147483647,
                   2147483647
               ],
               "avail_no_missing": [],
               "object_location_counts": [],
               "blocked_by": [
                   121
               ],
               "up_primary": -1,
               "acting_primary": 96,
               "purged_snaps": []
           },
           "empty": 0,
           "dne": 0,
           "incomplete": 0,
           "last_epoch_started": 616455,
           "hit_set_history": {
               "current_last_update": "0'0",
               "history": []
           }
       },
       "peer_info": [],
       "recovery_state": [
           {
               "name": "Started/Primary/Peering/Down",
               "enter_time": "2022-03-28 14:15:42.025185",
               "comment": "not enough up instances of this PG to go
active"
           },
           {
               "name": "Started/Primary/Peering",
               "enter_time": "2022-03-28 14:15:42.025128",
               "past_intervals": [
                   {
                       "first": "595151",
                       "last": "617660",
                       "all_participants": [
                           {
                               "osd": 77,
                               "shard": 0
                           },
                           {
                               "osd": 82,
                               "shard": 2
                           },
                           {
                               "osd": 90,
                               "shard": 0
                           },
                           {
                               "osd": 91,
                               "shard": 4
                           },
                           {
                               "osd": 96,
                               "shard": 2
                           },
                           {
                               "osd": 121,
                               "shard": 3
                           },
                           {
                               "osd": 140,
                               "shard": 1
                           },
                           {
                               "osd": 159,
                               "shard": 4
                           }
                       ],
                       "intervals": [
                           {
                               "first": "610237",
                               "last": "611219",
                               "acting": "96(2),121(3),159(4)"
                           },
                           {
                               "first": "611604",
                               "last": "616452",
                               "acting": "121(3),140(1),159(4)"
                           },
                           {
                               "first": "616461",
                               "last": "616518",
                               "acting": "96(2),140(1),159(4)"
                           }
                       ]
                   }
               ],
               "probing_osds": [
                   "77(0)",
                   "82(2)",
                   "90(0)",
                   "91(4)",
                   "96(2)",
                   "140(1)",
                   "159(4)"
               ],
               "blocked": "peering is blocked due to down osds",
               "down_osds_we_would_probe": [
                   121
               ],
               "peering_blocked_by": [
                   {
                       "osd": 121,
                       "current_lost_at": 0,
                       "comment": "starting or marking this osd lost may
let us proceed"
                   }
               ]
           },
           {
               "name": "Started",
               "enter_time": "2022-03-28 14:15:42.025068"
           }
       ],
       "agent_state": {}
}
[



--
Fulvio Galeazzi
GARR-CSD Department
skype: fgaleazzi70
tel.: +39-334-6533-250
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx

--
Fulvio Galeazzi
GARR-CSD Department
skype: fgaleazzi70
tel.: +39-334-6533-250
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux