Re: snapshots

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Fri, 15 Sep 2017, Wyllys Ingersoll wrote:
> Following up on this issue:
> 
> Does having too many cephfs snapshots prevent the deletion of older
> ones? We've got about 4800 snapshots and attempted to purge some older
> ones, but the "rmdir" of the old snapshots seems to hang and is not
> ever completing.
> 
> ceph -s shows an ever increasing number of blocked requests while this
> is going on. I waited over 10 minutes for 1 snapshot to delete before
> I finally killed the "find" command that was attempting to remove the
> old snapshots.  The blocked request count keeps going up anyway and
> the .snap directory seems unresponsive to 'ls' or other commands.

Sounds like a bug; it should be a quick operation.  What version?  Can you 
submit a bug report?  Preferably with a matching MDS log (and debug mds = 
20, debug ms = 1).

sage

> 
> 
> 
> On Fri, Sep 15, 2017 at 2:59 PM, Eric Eastman
> <eric.eastman@xxxxxxxxxxxxxx> wrote:
> > Just watch how many snapshots you create.  We hit an issues around
> > 4,700 snapshots of a singe file system.  See:
> > https://www.spinics.net/lists/ceph-devel/msg38203.html
> >
> > Eric
> >
> > On Fri, Sep 15, 2017 at 12:25 PM, Sage Weil <sage@xxxxxxxxxxxx> wrote:
> >> On Fri, 15 Sep 2017, Two Spirit wrote:
> >>> Excellent. Yes  was takling about CephFS. So basically it tags and
> >>> versions all objects and their corresponding metadata. Doesn't
> >>> actually replicate anything, and keeps the old objects around in the
> >>> OSDs?
> >>
> >> Something like that.  :)
> >>
> >> s
> >>
> >>>
> >>> On Fri, Sep 15, 2017 at 10:37 AM, Sage Weil <sage@xxxxxxxxxxxx> wrote:
> >>> > On Fri, 15 Sep 2017, Two Spirit wrote:
> >>> >> few questions
> >>> >
> >>> > I'm assuming you're talking about CephFS snapshots here:
> >>> >
> >>> >> 1) Is the idea of snapshotting about 100TB of data doable with Ceph?
> >>> >
> >>> > You can snapshot the entire file system (petabytes) if you like.
> >>> >
> >>> >> 2) How long would such a task take? Are we talking seconds, minutes,
> >>> >> hours, days, weeks?
> >>> >
> >>> > Seconds (if that).
> >>> >
> >>> >> 3) what does "stop i/o" mean. Does the filesystem unavailable to the
> >>> >> user during the snapshot time, or is this administratively unavailable
> >>> >> to the user, or does ceph stop io somewhere while in parallel the data
> >>> >> can be read/written to?
> >>> >
> >>> > There's no IO stoppage.  Clients essentially mark a barrier in their
> >>> > writeback caches so that previously buffered writes are contained in the
> >>> > snapshot and new writes are not.  How long it takes for that data to be
> >>> > flushed and stable/durable on OSDs depends on how big your client caches
> >>> > are, but that does not cause any io stoppage or gap in availability for
> >>> > users of the file system.
> >>> >
> >>> > sage
> >>> --
> >>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> >>> the body of a message to majordomo@xxxxxxxxxxxxxxx
> >>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >>>
> >>>
> >> --
> >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> >> the body of a message to majordomo@xxxxxxxxxxxxxxx
> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> > --
> > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> > the body of a message to majordomo@xxxxxxxxxxxxxxx
> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
> 
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux