I am running 0.87, In the end I just wiped the cluster and started again - it was quicker. Warren -----Original Message----- From: Gregory Farnum [mailto:greg@xxxxxxxxxxx] Sent: 12 February 2015 16:25 To: Jeffs, Warren (STFC,RAL,ISIS) Cc: ceph-users@xxxxxxxxxxxxxx Subject: Re: CephFS removal. What version of Ceph are you running? It's varied by a bit. But I think you want to just turn off the MDS and run the "fail" command — deactivate is actually the command for removing a logical MDS from the cluster, and you can't do that for a lone MDS because there's nobody to pass off the data to. I'll make a ticket to clarify this. When you've done that you should be able to delete it. -Greg On Mon, Feb 2, 2015 at 1:40 AM, <warren.jeffs@xxxxxxxxxx> wrote: > Hi All, > > > > Having a few problems removing cephfs file systems. > > > > I want to remove my current pools (was used for test data) – wiping > all current data, and start a fresh file system on my current cluster. > > > > I have looked over the documentation but I can’t find anything on > this. I have an object store pool, Which I don’t want to remove – but > I’d like to remove the cephfs file system pools and remake them. > > > > > > My cephfs is called ‘data’. > > > > Running ceph fs delete data returns: Error EINVAL: all MDS daemons > must be inactive before removing filesystem > > > > To make an MDS inactive I believe the command is: ceph mds deactivate > 0 > > > > Which returns: telling mds.0 135.248.53.134:6809/16692 to deactivate > > > > Checking the status of the mds using: ceph mds stat returns: e105: > 1/1/0 up {0=node2=up:stopping} > > > > This has been sitting at this status for the whole weekend with no > change. I don’t have any clients connected currently. > > > > When trying to manually just remove the pools, it’s not allowed as > there is a cephfs file system on them. > > > > I’m happy that all of the failsafe’s to stop someone removing a pool > are all working correctly. > > > > If this is currently undoable. Is there a way to quickly wipe a cephfs > filesystem – using RM from a kernel client is really slow. > > > > Many thanks > > > > Warren Jeffs > > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com