Hi Fulvio, I don't think upmap will help -- that is used to remap where data should be "up", but your problem is more that the PG chunks are not going active due to the bug. What happens if you export one of the PG chunks then import it to another OSD -- does that chunk become active? -- dan On Tue, Mar 29, 2022 at 10:51 AM Fulvio Galeazzi <fulvio.galeazzi@xxxxxxx> wrote: > > Hallo again Dan, I am afraid I'd need a little more help, please... > > Current status is as follows. > > This is where I moved the chunk which was on osd.121: > ~]# ceph-objectstore-tool --data-path /var/lib/ceph/osd/cephpa1-176 > --no-mon-config --op list-pgs | grep ^85\.25 > 85.25s3 > > while other chunks are (server - osd.id): > === r2srv101.pa1.box.garr - 96 > 85.25s2 > === r2srv100.pa1.box.garr - 121 <-- down, chunk is on osd.176 > 85.25s3 > === r2srv100.pa1.box.garr - 159 > 85.25s1 > 85.25s4 > === r1-sto09.pa1.box.garr - 56 > 85.25s4 > === r1-sto09.pa1.box.garr - 64 > 85.25s0 > === r3srv15.pa1.box.garr - 140 > 85.25s1 > > Health detail shows that just one chunk can be found (if I understand > the output correctly): > > ~]# ceph health detail | grep 85\.25 > pg 85.25 is stuck stale for 5680.315732, current state > stale+down+remapped, last acting > [2147483647,2147483647,96,2147483647,2147483647] > > Can I run some magic upmap command to explain my cluster where all the > chunks are? What would be the right syntax? > Little additional problem: I see s1 and s4 twice... I guess this was due > to remapping, as I was adding disks to the cluster: which one is the > right copy? > > Thanks! > > Fulvio > > Il 3/29/2022 9:35 AM, Fulvio Galeazzi ha scritto: > > Thanks a lot, Dan! > > > > > The EC pgs have a naming convention like 85.25s1 etc.. for the various > > > k/m EC shards. > > > > That was the bit of information I was missing... I was looking for the > > wrong object. > > I can now go on and export/import that one PGid chunk. > > > > Thanks again! > > > > Fulvio > > > > On 28/03/2022 16:27, Dan van der Ster wrote: > >> Hi Fulvio, > >> > >> You can check (offline) which PGs are on an OSD with the list-pgs op, > >> e.g. > >> > >> ceph-objectstore-tool --data-path /var/lib/ceph/osd/cephpa1-158/ > >> --op list-pgs > >> > >> > >> -- dan > >> > >> > >> On Mon, Mar 28, 2022 at 2:29 PM Fulvio Galeazzi > >> <fulvio.galeazzi@xxxxxxx> wrote: > >>> > >>> Hallo, > >>> all of a sudden, 3 of my OSDs failed, showing similar messages in > >>> the log: > >>> > >>> ..... > >>> -5> 2022-03-28 14:19:02.451 7fc20fe99700 5 osd.145 pg_epoch: > >>> 616454 pg[70.2c6s1( empty local-lis/les=612106/612107 n=0 > >>> ec=148456/148456 lis/c > >>> 612106/612106 les/c/f 612107/612107/0 612106/612106/612101) > >>> [168,145,102,96,112,124,128,134,56,34]p168(0) r=1 lpr=616429 crt=0'0 > >>> unknown mbc={}] > >>> enter Started > >>> -4> 2022-03-28 14:19:02.451 7fc20fe99700 5 osd.145 pg_epoch: > >>> 616454 pg[70.2c6s1( empty local-lis/les=612106/612107 n=0 > >>> ec=148456/148456 lis/c > >>> 612106/612106 les/c/f 612107/612107/0 612106/612106/612101) > >>> [168,145,102,96,112,124,128,134,56,34]p168(0) r=1 lpr=616429 crt=0'0 > >>> unknown mbc={}] > >>> enter Start > >>> -3> 2022-03-28 14:19:02.451 7fc20fe99700 1 osd.145 pg_epoch: > >>> 616454 pg[70.2c6s1( empty local-lis/les=612106/612107 n=0 > >>> ec=148456/148456 lis/c > >>> 612106/612106 les/c/f 612107/612107/0 612106/612106/612101) > >>> [168,145,102,96,112,124,128,134,56,34]p168(0) r=1 lpr=616429 crt=0'0 > >>> unknown mbc={}] > >>> state<Start>: transitioning to Stray > >>> -2> 2022-03-28 14:19:02.451 7fc20fe99700 5 osd.145 pg_epoch: > >>> 616454 pg[70.2c6s1( empty local-lis/les=612106/612107 n=0 > >>> ec=148456/148456 lis/c > >>> 612106/612106 les/c/f 612107/612107/0 612106/612106/612101) > >>> [168,145,102,96,112,124,128,134,56,34]p168(0) r=1 lpr=616429 crt=0'0 > >>> unknown mbc={}] > >>> exit Start 0.000008 0 0.000000 > >>> -1> 2022-03-28 14:19:02.451 7fc20fe99700 5 osd.145 pg_epoch: > >>> 616454 pg[70.2c6s1( empty local-lis/les=612106/612107 n=0 > >>> ec=148456/148456 lis/c > >>> 612106/612106 les/c/f 612107/612107/0 612106/612106/612101) > >>> [168,145,102,96,112,124,128,134,56,34]p168(0) r=1 lpr=616429 crt=0'0 > >>> unknown mbc={}] > >>> enter Started/Stray > >>> 0> 2022-03-28 14:19:02.451 7fc20f698700 -1 *** Caught signal > >>> (Aborted) ** > >>> in thread 7fc20f698700 thread_name:tp_osd_tp > >>> > >>> ceph version 14.2.22 (ca74598065096e6fcbd8433c8779a2be0c889351) > >>> nautilus (stable) > >>> 1: (()+0x12ce0) [0x7fc2327dcce0] > >>> 2: (gsignal()+0x10f) [0x7fc231452a4f] > >>> 3: (abort()+0x127) [0x7fc231425db5] > >>> 4: (ceph::__ceph_abort(char const*, int, char const*, > >>> std::__cxx11::basic_string<char, std::char_traits<char>, > >>> std::allocator<char> > const&)+0x1b4) [0x55b8139cb671] > >>> 5: (PG::check_past_interval_bounds() const+0xc16) [0x55b813b586f6] > >>> 6: (PG::RecoveryState::Reset::react(PG::AdvMap const&)+0x3e8) > >>> [0x55b813b963d8] > >>> 7: (boost::statechart::simple_state<PG::RecoveryState::Reset, > >>> PG::RecoveryState::RecoveryMachine, boost::mpl::list<mpl_::na, mpl_::na, > >>> mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, > >>> mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, > >>> mpl_::na, mpl_::na, mpl_::na, mpl_::na>, > >>> (boost::statechart::history_mode)0>::react_impl(boost::statechart::event_base > >>> > >>> const&, void const*)+0x7d) [0x55b813bdd32d] > >>> 8: (PG::handle_advance_map(std::shared_ptr<OSDMap const>, > >>> std::shared_ptr<OSDMap const>, std::vector<int, std::allocator<int> >&, > >>> int, std::vector<int, std::allocator<int> >&, int, > >>> PG::RecoveryCtx*)+0x39d) [0x55b813b7b5fd] > >>> 9: (OSD::advance_pg(unsigned int, PG*, ThreadPool::TPHandle&, > >>> PG::RecoveryCtx*)+0x2e9) [0x55b813ad14e9] > >>> 10: (OSD::dequeue_peering_evt(OSDShard*, PG*, > >>> std::shared_ptr<PGPeeringEvent>, ThreadPool::TPHandle&)+0xaa) > >>> [0x55b813ae345a] > >>> 11: (PGPeeringItem::run(OSD*, OSDShard*, boost::intrusive_ptr<PG>&, > >>> ThreadPool::TPHandle&)+0x55) [0x55b813d66c15] > >>> 12: (OSD::ShardedOpWQ::_process(unsigned int, > >>> ceph::heartbeat_handle_d*)+0x1366) [0x55b813adff46] > >>> 13: (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x5c4) > >>> [0x55b8140dc944] > >>> 14: (ShardedThreadPool::WorkThreadSharded::entry()+0x14) > >>> [0x55b8140df514] > >>> 15: (()+0x81cf) [0x7fc2327d21cf] > >>> 16: (clone()+0x43) [0x7fc23143dd83] > >>> > >>> Trying to "activate --all", rebotting server, and such, did not help. > >>> > >>> I am now stuck with one PG (85.25) down, find below the output from > >>> "query". > >>> > >>> The PG belongs to a 3+2 erasure-coded pool. > >>> As the devices corresponding to the 3 down OSDs are properly mounted, is > >>> there a way to get PG.ID=85.25 from the devices and copy it elsewhere? > >>> Actually, I tried to find 85.25 in the 3 down OSDs with command: > >>> ~]# ceph-objectstore-tool --data-path /var/lib/ceph/osd/cephpa1-158/ > >>> --no-mon-config --pgid 85.25 --op export --file /tmp/pg_85-25 > >>> PG '85.25' not found > >>> which puzzled me... is there a way to search such PG.ID over the > >>> whole cluster? > >>> > >>> Thanks for your help! > >>> > >>> Fulvio > >>> > >>> ==================== > >>> > >>> ~]# ceph --cluster cephpa1 health detail | grep down > >>> ..... > >>> PG_AVAILABILITY Reduced data availability: 1 pg inactive, 1 pg down > >>> pg 85.25 is down+remapped, acting > >>> [2147483647,2147483647,96,2147483647,2147483647] > >>> ~]# ceph --cluster cephpa1 pg 85.25 query > >>> { > >>> "state": "down+remapped", > >>> "snap_trimq": "[]", > >>> "snap_trimq_len": 0, > >>> "epoch": 617667, > >>> "up": [ > >>> 2147483647, > >>> 2147483647, > >>> 2147483647, > >>> 2147483647, > >>> 2147483647 > >>> ], > >>> "acting": [ > >>> 2147483647, > >>> 2147483647, > >>> 96, > >>> 2147483647, > >>> 2147483647 > >>> ], > >>> "info": { > >>> "pgid": "85.25s2", > >>> "last_update": "606021'521273", > >>> "last_complete": "606021'521273", > >>> "log_tail": "605873'518175", > >>> "last_user_version": 521273, > >>> "last_backfill": "MAX", > >>> "last_backfill_bitwise": 1, > >>> "purged_snaps": [], > >>> "history": { > >>> "epoch_created": 478130, > >>> "epoch_pool_created": 478130, > >>> "last_epoch_started": 611605, > >>> "last_interval_started": 611604, > >>> "last_epoch_clean": 595152, > >>> "last_interval_clean": 595151, > >>> "last_epoch_split": 0, > >>> "last_epoch_marked_full": 0, > >>> "same_up_since": 617661, > >>> "same_interval_since": 617661, > >>> "same_primary_since": 616519, > >>> "last_scrub": "597862'517979", > >>> "last_scrub_stamp": "2022-03-15 05:08:42.561720", > >>> "last_deep_scrub": "597454'511650", > >>> "last_deep_scrub_stamp": "2022-03-09 18:05:34.205121", > >>> "last_clean_scrub_stamp": "2022-03-15 05:08:42.561720" > >>> }, > >>> "stats": { > >>> "version": "606021'521273", > >>> "reported_seq": "1733225", > >>> "reported_epoch": "617667", > >>> "state": "down+remapped", > >>> "last_fresh": "2022-03-28 14:17:48.920542", > >>> "last_change": "2022-03-28 14:15:42.025193", > >>> "last_active": "2022-03-25 15:24:42.557667", > >>> "last_peered": "2022-03-25 15:24:25.912029", > >>> "last_clean": "2022-03-15 01:50:02.979366", > >>> "last_became_active": "2022-03-25 12:44:52.379711", > >>> "last_became_peered": "2022-03-25 12:44:52.379711", > >>> "last_unstale": "2022-03-28 14:17:48.920542", > >>> "last_undegraded": "2022-03-28 14:17:48.920542", > >>> "last_fullsized": "2022-03-28 14:17:48.920542", > >>> "mapping_epoch": 617661, > >>> "log_start": "605873'518175", > >>> "ondisk_log_start": "605873'518175", > >>> "created": 478130, > >>> "last_epoch_clean": 595152, > >>> "parent": "0.0", > >>> "parent_split_bits": 0, > >>> "last_scrub": "597862'517979", > >>> "last_scrub_stamp": "2022-03-15 05:08:42.561720", > >>> "last_deep_scrub": "597454'511650", > >>> "last_deep_scrub_stamp": "2022-03-09 18:05:34.205121", > >>> "last_clean_scrub_stamp": "2022-03-15 05:08:42.561720", > >>> "log_size": 3098, > >>> "ondisk_log_size": 3098, > >>> "stats_invalid": false, > >>> "dirty_stats_invalid": false, > >>> "omap_stats_invalid": false, > >>> "hitset_stats_invalid": false, > >>> "hitset_bytes_stats_invalid": false, > >>> "pin_stats_invalid": false, > >>> "manifest_stats_invalid": false, > >>> "snaptrimq_len": 0, > >>> "stat_sum": { > >>> "num_bytes": 165108592640, > >>> "num_objects": 39402, > >>> "num_object_clones": 0, > >>> "num_object_copies": 197010, > >>> "num_objects_missing_on_primary": 0, > >>> "num_objects_missing": 0, > >>> "num_objects_degraded": 0, > >>> "num_objects_misplaced": 0, > >>> "num_objects_unfound": 0, > >>> "num_objects_dirty": 39402, > >>> "num_whiteouts": 0, > >>> "num_read": 338705, > >>> "num_read_kb": 124297874, > >>> "num_write": 316971, > >>> "num_write_kb": 311968402, > >>> "num_scrub_errors": 0, > >>> "num_shallow_scrub_errors": 0, > >>> "num_deep_scrub_errors": 0, > >>> "num_objects_recovered": 196778, > >>> "num_bytes_recovered": 824577241088, > >>> "num_keys_recovered": 0, > >>> "num_objects_omap": 0, > >>> "num_objects_hit_set_archive": 0, > >>> "num_bytes_hit_set_archive": 0, > >>> "num_flush": 0, > >>> "num_flush_kb": 0, > >>> "num_evict": 0, > >>> "num_evict_kb": 0, > >>> "num_promote": 0, > >>> "num_flush_mode_high": 0, > >>> "num_flush_mode_low": 0, > >>> "num_evict_mode_some": 0, > >>> "num_evict_mode_full": 0, > >>> "num_objects_pinned": 0, > >>> "num_legacy_snapsets": 0, > >>> "num_large_omap_objects": 0, > >>> "num_objects_manifest": 0, > >>> "num_omap_bytes": 0, > >>> "num_omap_keys": 0, > >>> "num_objects_repaired": 0 > >>> }, > >>> "up": [ > >>> 2147483647, > >>> 2147483647, > >>> 2147483647, > >>> 2147483647, > >>> 2147483647 > >>> ], > >>> "acting": [ > >>> 2147483647, > >>> 2147483647, > >>> 96, > >>> 2147483647, > >>> 2147483647 > >>> ], > >>> "avail_no_missing": [], > >>> "object_location_counts": [], > >>> "blocked_by": [ > >>> 121 > >>> ], > >>> "up_primary": -1, > >>> "acting_primary": 96, > >>> "purged_snaps": [] > >>> }, > >>> "empty": 0, > >>> "dne": 0, > >>> "incomplete": 0, > >>> "last_epoch_started": 616455, > >>> "hit_set_history": { > >>> "current_last_update": "0'0", > >>> "history": [] > >>> } > >>> }, > >>> "peer_info": [], > >>> "recovery_state": [ > >>> { > >>> "name": "Started/Primary/Peering/Down", > >>> "enter_time": "2022-03-28 14:15:42.025185", > >>> "comment": "not enough up instances of this PG to go > >>> active" > >>> }, > >>> { > >>> "name": "Started/Primary/Peering", > >>> "enter_time": "2022-03-28 14:15:42.025128", > >>> "past_intervals": [ > >>> { > >>> "first": "595151", > >>> "last": "617660", > >>> "all_participants": [ > >>> { > >>> "osd": 77, > >>> "shard": 0 > >>> }, > >>> { > >>> "osd": 82, > >>> "shard": 2 > >>> }, > >>> { > >>> "osd": 90, > >>> "shard": 0 > >>> }, > >>> { > >>> "osd": 91, > >>> "shard": 4 > >>> }, > >>> { > >>> "osd": 96, > >>> "shard": 2 > >>> }, > >>> { > >>> "osd": 121, > >>> "shard": 3 > >>> }, > >>> { > >>> "osd": 140, > >>> "shard": 1 > >>> }, > >>> { > >>> "osd": 159, > >>> "shard": 4 > >>> } > >>> ], > >>> "intervals": [ > >>> { > >>> "first": "610237", > >>> "last": "611219", > >>> "acting": "96(2),121(3),159(4)" > >>> }, > >>> { > >>> "first": "611604", > >>> "last": "616452", > >>> "acting": "121(3),140(1),159(4)" > >>> }, > >>> { > >>> "first": "616461", > >>> "last": "616518", > >>> "acting": "96(2),140(1),159(4)" > >>> } > >>> ] > >>> } > >>> ], > >>> "probing_osds": [ > >>> "77(0)", > >>> "82(2)", > >>> "90(0)", > >>> "91(4)", > >>> "96(2)", > >>> "140(1)", > >>> "159(4)" > >>> ], > >>> "blocked": "peering is blocked due to down osds", > >>> "down_osds_we_would_probe": [ > >>> 121 > >>> ], > >>> "peering_blocked_by": [ > >>> { > >>> "osd": 121, > >>> "current_lost_at": 0, > >>> "comment": "starting or marking this osd lost may > >>> let us proceed" > >>> } > >>> ] > >>> }, > >>> { > >>> "name": "Started", > >>> "enter_time": "2022-03-28 14:15:42.025068" > >>> } > >>> ], > >>> "agent_state": {} > >>> } > >>> [ > >>> > >>> > >>> > >>> -- > >>> Fulvio Galeazzi > >>> GARR-CSD Department > >>> skype: fgaleazzi70 > >>> tel.: +39-334-6533-250 > >>> _______________________________________________ > >>> ceph-users mailing list -- ceph-users@xxxxxxx > >>> To unsubscribe send an email to ceph-users-leave@xxxxxxx > > -- > Fulvio Galeazzi > GARR-CSD Department > skype: fgaleazzi70 > tel.: +39-334-6533-250 _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx