Deal all: I have a ceph cluster with 150 osds .Ceph version is 12.2.10 luminous. Pool vms. Few days ago . pool vms set size 3. And I make a backup for pg 10.1a4 which belong to pool vms. Use ceph-objectstore-tool export command Now I set pool vms size to 1 。 after rebalance. Pg 10.1a4 stay in stale + incomplete . I import the old pg data I have backuped. Use ceph-objectstore-tool import –pgid 10.1a4 But the osd crash. Here is the log -6> 2021-12-28 18:06:30.915130 7f4f517f6700 5 -- 10.222.3.24:6813/1339275 >> 10.222.3.22:0/3043732 conn(0x56431088b800 :6813 s=STATE_OPEN_MESSAGE_READ_FOOTER_AND_DISPATCH pgs=454 cs=1 l=1). rx osd.69 seq 1 0x564310831400 osd_ping(ping e117671 stamp 2021-12-28 18:06:30.914946) v4 -5> 2021-12-28 18:06:30.915141 7f4f51ff7700 5 -- 10.222.3.24:6815/1339275 >> 10.222.3.22:0/3043732 conn(0x56431088d000 :6815 s=STATE_OPEN_MESSAGE_READ_FOOTER_AND_DISPATCH pgs=454 cs=1 l=1). rx osd.69 seq 1 0x56431085ba00 osd_ping(ping e117671 stamp 2021-12-28 18:06:30.914946) v4 -4> 2021-12-28 18:06:30.915164 7f4f517f6700 1 -- 10.222.3.24:6813/1339275 <== osd.69 10.222.3.22:0/3043732 1 ==== osd_ping(ping e117671 stamp 2021-12-28 18:06:30.914946) v4 ==== 2004+0+0 (2371603292 0 0) 0x564310831400 con 0x56431088b800 -3> 2021-12-28 18:06:30.915187 7f4f51ff7700 1 -- 10.222.3.24:6815/1339275 <== osd.69 10.222.3.22:0/3043732 1 ==== osd_ping(ping e117671 stamp 2021-12-28 18:06:30.914946) v4 ==== 2004+0+0 (2371603292 0 0) 0x56431085ba00 con 0x56431088d000 -2> 2021-12-28 18:06:30.915196 7f4f517f6700 1 -- 10.222.3.24:6813/1339275 --> 10.222.3.22:0/3043732 -- osd_ping(ping_reply e117671 stamp 2021-12-28 18:06:30.914946) v4 -- 0x564310803a00 con 0 -1> 2021-12-28 18:06:30.915236 7f4f51ff7700 1 -- 10.222.3.24:6815/1339275 --> 10.222.3.22:0/3043732 -- osd_ping(ping_reply e117671 stamp 2021-12-28 18:06:30.914946) v4 -- 0x564310956000 con 0 0> 2021-12-28 18:06:31.003766 7f4f39ffd700 -1 /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MACHINE_SIZE/huge/release/12.2.11/rpm/el7/BUILD/ceph-12.2.11/src/common/Throttle.cc: In function 'int64_t Throttle::take(int64_t)' thread 7f4f39ffd700 time 2021-12-28 18:06:30.938147 /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MACHINE_SIZE/huge/release/12.2.11/rpm/el7/BUILD/ceph-12.2.11/src/common/Throttle.cc: 148: FAILED assert(c >= 0) ceph version 12.2.11 (26dc3775efc7bb286a1d6d66faee0ba30ea23eee) luminous (stable) 1: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x110) [0x5642fb767120] 2: (Throttle::take(long)+0x2a9) [0x5642fb75cc49] 3: (Objecter::_op_submit_with_budget(Objecter::Op*, ceph::shunique_lock<boost::shared_mutex>&, unsigned long*, int*)+0x2a2) [0x5642fb471b12] 4: (Objecter::op_submit(Objecter::Op*, unsigned long*, int*)+0x7a) [0x5642fb471d2a] 5: (PrimaryLogPG::_copy_some(std::shared_ptr<ObjectContext>, std::shared_ptr<PrimaryLogPG::CopyOp>)+0xf75) [0x5642fb325625] 6: (PrimaryLogPG::start_copy(PrimaryLogPG::CopyCallback*, std::shared_ptr<ObjectContext>, hobject_t, object_locator_t, unsigned long, unsigned int, bool, unsigned int, unsigned int)+0x6e4) [0x5642fb3265f4] 7: (PrimaryLogPG::do_osd_ops(PrimaryLogPG::OpContext*, std::vector<OSDOp, std::allocator<OSDOp> >&)+0x7e8e) [0x5642fb36fa1e] 8: (PrimaryLogPG::prepare_transaction(PrimaryLogPG::OpContext*)+0xbf) [0x5642fb379a9f] 9: (PrimaryLogPG::execute_ctx(PrimaryLogPG::OpContext*)+0x753) [0x5642fb37a823] 10: (PrimaryLogPG::do_op(boost::intrusive_ptr<OpRequest>&)+0x3147) [0x5642fb37f207] 11: (PrimaryLogPG::do_request(boost::intrusive_ptr<OpRequest>&, ThreadPool::TPHandle&)+0xebb) [0x5642fb33c1db] 12: (OSD::dequeue_op(boost::intrusive_ptr<PG>, boost::intrusive_ptr<OpRequest>, ThreadPool::TPHandle&)+0x3f9) [0x5642fb1b84f9] 13: (PGQueueable::RunVis::operator()(boost::intrusive_ptr<OpRequest> const&)+0x57) [0x5642fb446957] 14: (OSD::ShardedOpWQ::_process(unsigned int, ceph::heartbeat_handle_d*)+0xfbe) [0x5642fb1e778e] 15: (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x839) [0x5642fb76cc39] 16: (ShardedThreadPool::WorkThreadSharded::entry()+0x10) [0x5642fb76ebd0] 17: (()+0x7dd5) [0x7f4f550d9dd5] 18: (clone()+0x6d) [0x7f4f541c9ead] NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this. How can I import the old data. If I allow lost pg data . how can I create a new pg with active + clean. Look forward to your reply. _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx