On Fri, 2 Feb 2018, Frank Li wrote: > Hi, I ran the ceph osd force-create-pg command in luminious 12.2.2 to recover a failed pg, and it > Instantly caused all of the monitor to crash, is there anyway to revert back to an earlier state of the cluster ? > Right now, the monitors refuse to come up, the error message is as follows: > I’ve filed a ceph ticket for the crash, but just wonder if there is a way to get the cluster back up ? > > https://tracker.ceph.com/issues/22847 Can you includ the bit of the log a few lines up that includes the assertion and file line number that failed? Also, "during the course of trouble-shooting an osd issue" makes me nervous: force-create-pg creates a new, *empty* PG when all copies of the old one have been lost. Is that what you meant to do? It is essentially telling the system to give up and accepting that there is data loss. Is that what you meant? Thanks! sage > > --- begin dump of recent events --- > 0> 2018-01-31 22:47:22.959665 7fc64350e700 -1 *** Caught signal (Aborted) ** > in thread 7fc64350e700 thread_name:cpu_tp > > ceph version 12.2.2 (cf0baeeeeba3b47f9427c6c97e2144b094b7e5ba) luminous (stable) > 1: (()+0x8eae11) [0x55f1113fae11] > 2: (()+0xf5e0) [0x7fc64aafa5e0] > 3: (gsignal()+0x37) [0x7fc647fca1f7] > 4: (abort()+0x148) [0x7fc647fcb8e8] > 5: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x284) [0x55f1110fa4a4] > 6: (()+0x2ccc4e) [0x55f110ddcc4e] > 7: (OSDMonitor::update_creating_pgs()+0x98b) [0x55f11102232b] > 8: (C_UpdateCreatingPGs::finish(int)+0x79) [0x55f1110777b9] > 9: (Context::complete(int)+0x9) [0x55f110ed30c9] > 10: (ParallelPGMapper::WQ::_process(ParallelPGMapper::Item*, ThreadPool::TPHandle&)+0x7f) [0x55f111204e1f] > 11: (ThreadPool::worker(ThreadPool::WorkThread*)+0xa8e) [0x55f111100f1e] > 12: (ThreadPool::WorkThread::entry()+0x10) [0x55f111101e00] > 13: (()+0x7e25) [0x7fc64aaf2e25] > 14: (clone()+0x6d) [0x7fc64808d34d] > NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this. > > -- > Efficiency is Intelligent Laziness >
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com