Re: Cinder + CEPH Storage Full Scenario

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Sorry about that, I guess newer releases than my Dumpling calculate it differently, then.
I can take a look tomorrow at the exact numbers I get, but I'm pretty sure it's just a sum on D.

Jan

> On 19 Oct 2015, at 20:40, John Spray <jspray@xxxxxxxxxx> wrote:
> 
> On Mon, Oct 19, 2015 at 7:28 PM, Jan Schermer <jan@xxxxxxxxxxx> wrote:
>> Cinder checking free space will not help.
>> You will get one full OSD long before you run "out of space" from Ceph
>> perspective, and it gets worse with the number of OSDs you have. Using 99%
>> of space in Ceph is not the same as having all the OSDs 99% full because the
>> data is not distributed in a completely fair fashion. Not sure how much that
>> can be helped, but my cluster can store at most 2TB of data when claiming to
>> have 14TB free.
> 
> The "max_avail" per-pool number that you get out of "ceph df" is aware
> of this, and will calculate that actual writeable capacity based on
> whatever OSD has the least available space.  From a quick look at the
> code, it seems that the RBD Cinder plugin reports free_capacity_gb
> from max_avail, so unless you're seeing a different behaviour I don't
> think we have a problem.
> 
> This is me looking at master ceph and master cinder, so no idea which
> released versions got this behaviour (the cinder code was modified in
> March this year).
> 
> John
> 
>> 
>> You *really* need to monitor each OSD's free space and treat it with utmost
>> criticality.
>> 
>> Jan
>> 
>> 
>> On 19 Oct 2015, at 20:00, Andrew Woodward <xarses@xxxxxxxxx> wrote:
>> 
>> Cinder will periodically inspect the free space of the volume services and
>> use this data when determining which one to schedule to when a request is
>> received. In this case the cinder volume create request may error out in
>> scheduling. You may also see an error when instantiating a volume from an
>> image if it passes the prior but then becomes out of space during writing
>> the image to the volume.
>> 
>> I'm not sure if it's still the case, but in Havana (I see no reason for it
>> to change) the free space check in cinder didn't account for the difference
>> between promised space (the max of the volumes assigned) instead it would
>> literally look for free space in the output of `rados df`
>> 
>> As noted above if the cluster gets to "100%" used, bad things will happen to
>> your VM's. The most likely case is that they all assert read-only
>> filesystems. (100% is a missnomer as there is a configured max % where it
>> will stop accepting data writes to ensure that important object replication
>> / maintenance can occur and have the cluster not fall over)
>> 
>> On Mon, Oct 19, 2015 at 7:51 AM LOPEZ Jean-Charles <jelopez@xxxxxxxxxx>
>> wrote:
>>> 
>>> Hi,
>>> 
>>> when an OSD gets full, any write operation to the entire cluster will be
>>> disabled.
>>> 
>>> As a result, creating a single RBD will become impossible and all VMs that
>>> need to write to one of their Ceph back RBDs will suffer the same pain.
>>> 
>>> Usually, this ends up as a bad sorry for the VMs.
>>> 
>>> The best practice is to monitor the disk space usage for the OSDs and as a
>>> matter of fact RHCS 1.# includes a cep old df command to do this. You can
>>> also use the output of the cep old report command to grab the appropriate
>>> info to compute it or rely on external SNMP monitoring tools to grab the
>>> usage details of the particular OSD disk drives.
>>> 
>>> Have a great day.
>>> JC
>>> 
>>>> On Oct 19, 2015, at 02:32, Bharath Krishna <BKrishna@xxxxxxxxxxxxxxx>
>>>> wrote:
>>>> 
>>>> I mean cluster OSDs are physically full.
>>>> 
>>>> I understand its not a pretty way to operate CEPH allowing to become
>>>> full,
>>>> but I just wanted to know the boundary condition if it becomes full.
>>>> 
>>>> Will cinder create volume operation creates new volume at all or error
>>>> is
>>>> thrown at Cinder API level itself stating that no space available?
>>>> 
>>>> When IO stalls, will I be able to read the data from CEPH cluster I.e
>>>> can
>>>> I still read data from existing volumes created from CEPH cluster?
>>>> 
>>>> Thanks for the quick reply.
>>>> 
>>>> Regards
>>>> M Bharath Krishna
>>>> 
>>>> On 10/19/15, 2:51 PM, "Jan Schermer" <jan@xxxxxxxxxxx> wrote:
>>>> 
>>>>> Do you mean when the CEPH cluster (OSDs) is physically full or when the
>>>>> quota is reached?
>>>>> 
>>>>> If CEPH becomes full it just stalls all IO (maybe just write IO, but
>>>>> effectively same thing) - not pretty and you must never ever let it
>>>>> become full.
>>>>> 
>>>>> Jan
>>>>> 
>>>>> 
>>>>>> On 19 Oct 2015, at 11:15, Bharath Krishna <BKrishna@xxxxxxxxxxxxxxx>
>>>>>> wrote:
>>>>>> 
>>>>>> Hi
>>>>>> 
>>>>>> What happens when Cinder service with CEPH backend storage cluster
>>>>>> capacity is FULL?
>>>>>> 
>>>>>> What would be the out come of new cinder create volume request?
>>>>>> 
>>>>>> Will volume be created with space not available for use or an error
>>>>>> thrown from Cinder API stating no space available for new volume.
>>>>>> 
>>>>>> I could not try this in my environment and fill up the cluster.
>>>>>> 
>>>>>> Please reply if you have ever tried and tested this.
>>>>>> 
>>>>>> Thank you.
>>>>>> 
>>>>>> Regards,
>>>>>> M Bharath Krishna
>>>>>> _______________________________________________
>>>>>> ceph-users mailing list
>>>>>> ceph-users@xxxxxxxxxxxxxx
>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>> 
>>>> 
>>>> _______________________________________________
>>>> ceph-users mailing list
>>>> ceph-users@xxxxxxxxxxxxxx
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>> 
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@xxxxxxxxxxxxxx
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> 
>> --
>> 
>> --
>> 
>> Andrew Woodward
>> 
>> Mirantis
>> 
>> Fuel Community Ambassador
>> 
>> Ceph Community
>> 
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@xxxxxxxxxxxxxx
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> 
>> 
>> 
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@xxxxxxxxxxxxxx
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> 

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux