Recovery pg from backup

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Deal all:
    I  have a ceph cluster with 150 osds .Ceph version is 12.2.10  luminous. Pool vms.


Few days ago . pool vms set size 3. And I  make a backup for pg 10.1a4 which belong to pool vms. Use ceph-objectstore-tool export command

Now I set pool vms size to 1 。 after rebalance.
Pg  10.1a4 stay in  stale + incomplete .

I import the old pg data  I have backuped.  Use ceph-objectstore-tool import –pgid 10.1a4
But the osd crash.

Here is the log

    -6> 2021-12-28 18:06:30.915130 7f4f517f6700  5 -- 10.222.3.24:6813/1339275 >> 10.222.3.22:0/3043732 conn(0x56431088b800 :6813 s=STATE_OPEN_MESSAGE_READ_FOOTER_AND_DISPATCH pgs=454 cs=1 l=1). rx osd.69 seq 1 0x564310831400 osd_ping(ping e117671 stamp 2021-12-28 18:06:30.914946) v4
    -5> 2021-12-28 18:06:30.915141 7f4f51ff7700  5 -- 10.222.3.24:6815/1339275 >> 10.222.3.22:0/3043732 conn(0x56431088d000 :6815 s=STATE_OPEN_MESSAGE_READ_FOOTER_AND_DISPATCH pgs=454 cs=1 l=1). rx osd.69 seq 1 0x56431085ba00 osd_ping(ping e117671 stamp 2021-12-28 18:06:30.914946) v4
    -4> 2021-12-28 18:06:30.915164 7f4f517f6700  1 -- 10.222.3.24:6813/1339275 <== osd.69 10.222.3.22:0/3043732 1 ==== osd_ping(ping e117671 stamp 2021-12-28 18:06:30.914946) v4 ==== 2004+0+0 (2371603292 0 0) 0x564310831400 con 0x56431088b800
    -3> 2021-12-28 18:06:30.915187 7f4f51ff7700  1 -- 10.222.3.24:6815/1339275 <== osd.69 10.222.3.22:0/3043732 1 ==== osd_ping(ping e117671 stamp 2021-12-28 18:06:30.914946) v4 ==== 2004+0+0 (2371603292 0 0) 0x56431085ba00 con 0x56431088d000
    -2> 2021-12-28 18:06:30.915196 7f4f517f6700  1 -- 10.222.3.24:6813/1339275 --> 10.222.3.22:0/3043732 -- osd_ping(ping_reply e117671 stamp 2021-12-28 18:06:30.914946) v4 -- 0x564310803a00 con 0
    -1> 2021-12-28 18:06:30.915236 7f4f51ff7700  1 -- 10.222.3.24:6815/1339275 --> 10.222.3.22:0/3043732 -- osd_ping(ping_reply e117671 stamp 2021-12-28 18:06:30.914946) v4 -- 0x564310956000 con 0
     0> 2021-12-28 18:06:31.003766 7f4f39ffd700 -1 /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MACHINE_SIZE/huge/release/12.2.11/rpm/el7/BUILD/ceph-12.2.11/src/common/Throttle.cc: In function 'int64_t Throttle::take(int64_t)' thread 7f4f39ffd700 time 2021-12-28 18:06:30.938147
/home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MACHINE_SIZE/huge/release/12.2.11/rpm/el7/BUILD/ceph-12.2.11/src/common/Throttle.cc: 148: FAILED assert(c >= 0)

ceph version 12.2.11 (26dc3775efc7bb286a1d6d66faee0ba30ea23eee) luminous (stable)
1: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x110) [0x5642fb767120]
2: (Throttle::take(long)+0x2a9) [0x5642fb75cc49]
3: (Objecter::_op_submit_with_budget(Objecter::Op*, ceph::shunique_lock<boost::shared_mutex>&, unsigned long*, int*)+0x2a2) [0x5642fb471b12]
4: (Objecter::op_submit(Objecter::Op*, unsigned long*, int*)+0x7a) [0x5642fb471d2a]
5: (PrimaryLogPG::_copy_some(std::shared_ptr<ObjectContext>, std::shared_ptr<PrimaryLogPG::CopyOp>)+0xf75) [0x5642fb325625]
6: (PrimaryLogPG::start_copy(PrimaryLogPG::CopyCallback*, std::shared_ptr<ObjectContext>, hobject_t, object_locator_t, unsigned long, unsigned int, bool, unsigned int, unsigned int)+0x6e4) [0x5642fb3265f4]
7: (PrimaryLogPG::do_osd_ops(PrimaryLogPG::OpContext*, std::vector<OSDOp, std::allocator<OSDOp> >&)+0x7e8e) [0x5642fb36fa1e]
8: (PrimaryLogPG::prepare_transaction(PrimaryLogPG::OpContext*)+0xbf) [0x5642fb379a9f]
9: (PrimaryLogPG::execute_ctx(PrimaryLogPG::OpContext*)+0x753) [0x5642fb37a823]
10: (PrimaryLogPG::do_op(boost::intrusive_ptr<OpRequest>&)+0x3147) [0x5642fb37f207]
11: (PrimaryLogPG::do_request(boost::intrusive_ptr<OpRequest>&, ThreadPool::TPHandle&)+0xebb) [0x5642fb33c1db]
12: (OSD::dequeue_op(boost::intrusive_ptr<PG>, boost::intrusive_ptr<OpRequest>, ThreadPool::TPHandle&)+0x3f9) [0x5642fb1b84f9]
13: (PGQueueable::RunVis::operator()(boost::intrusive_ptr<OpRequest> const&)+0x57) [0x5642fb446957]
14: (OSD::ShardedOpWQ::_process(unsigned int, ceph::heartbeat_handle_d*)+0xfbe) [0x5642fb1e778e]
15: (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x839) [0x5642fb76cc39]
16: (ShardedThreadPool::WorkThreadSharded::entry()+0x10) [0x5642fb76ebd0]
17: (()+0x7dd5) [0x7f4f550d9dd5]
18: (clone()+0x6d) [0x7f4f541c9ead]
NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this.


How can I import the old data.
If I allow lost pg data . how can I create a new pg with active + clean.

Look forward to your reply.
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux