Re: Luminous cluster in very bad state need some assistance.

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Mon, 4 Feb 2019, Philippe Van Hecke wrote:
> Hi Sage,
> 
> I try to make the following.
> 
> ceph-objectstore-tool --data-path /var/lib/ceph/osd/ceph-49/ --journal /var/lib/ceph/osd/ceph-49/journal --pgid 11.182 --op export-remove --debug --file /tmp/export-pg/18.182 2>ceph-objectstore-tool-export-remove.txt
> but this rise exception 
> 
> find here  https://filesender.belnet.be/?s=download&token=e2b1fdbc-0739-423f-9d97-0bd258843a33 file ceph-objectstore-tool-export-remove.txt

In that case,  cp --preserve=all 
/var/lib/ceph/osd/ceph-49/current/11.182_head to a safe location and then 
use the ceph-objecstore-tool --op remove command.  But first confirm that 
'ceph pg ls' shows the PG as active.

sage


 > > Kr
> 
> Philippe.
> 
> ________________________________________
> From: Sage Weil <sage@xxxxxxxxxxxx>
> Sent: 04 February 2019 06:59
> To: Philippe Van Hecke
> Cc: ceph-users@xxxxxxxxxxxxxx; Belnet Services
> Subject: Re:  Luminous cluster in very bad state need some assistance.
> 
> On Mon, 4 Feb 2019, Philippe Van Hecke wrote:
> > Hi Sage, First of all tanks for your help
> >
> > Please find here  https://filesender.belnet.be/?s=download&token=dea0edda-5b6a-4284-9ea1-c1fdf88b65e9
> > the osd log with debug info for osd.49. and indeed if all buggy osd can restart that can may be solve the issue.
> > But i also happy that you confirm my understanding that in the worst case removing pool can also resolve the problem even in this case i lose data  but finish with a working cluster.
> 
> If PGs are damaged, removing the pool would be part of getting to
> HEALTH_OK, but you'd probably also need to remove any problematic PGs that
> are preventing the OSD starting.
> 
> But keep in mind that (1) i see 3 PGs that don't peer spread across pools
> 11 and 12; not sure which one you are considering deleting.  Also (2) if
> one pool isn't fully available it generall won't be a problem for other
> pools, as long as the osds start.  And doing ceph-objectstore-tool
> export-remove is a pretty safe way to move any problem PGs out of the way
> to get your OSDs starting--just make sure you hold onto that backup/export
> because you may need it later!
> 
> > PS: don't know and don't want to open debat about top/bottom posting but would like to know the preference of this list :-)
> 
> No preference :)
> 
> sage
> 
> 
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux