On Mon, Oct 19, 2015 at 7:28 PM, Jan Schermer <jan@xxxxxxxxxxx> wrote: > Cinder checking free space will not help. > You will get one full OSD long before you run "out of space" from Ceph > perspective, and it gets worse with the number of OSDs you have. Using 99% > of space in Ceph is not the same as having all the OSDs 99% full because the > data is not distributed in a completely fair fashion. Not sure how much that > can be helped, but my cluster can store at most 2TB of data when claiming to > have 14TB free. The "max_avail" per-pool number that you get out of "ceph df" is aware of this, and will calculate that actual writeable capacity based on whatever OSD has the least available space. From a quick look at the code, it seems that the RBD Cinder plugin reports free_capacity_gb from max_avail, so unless you're seeing a different behaviour I don't think we have a problem. This is me looking at master ceph and master cinder, so no idea which released versions got this behaviour (the cinder code was modified in March this year). John > > You *really* need to monitor each OSD's free space and treat it with utmost > criticality. > > Jan > > > On 19 Oct 2015, at 20:00, Andrew Woodward <xarses@xxxxxxxxx> wrote: > > Cinder will periodically inspect the free space of the volume services and > use this data when determining which one to schedule to when a request is > received. In this case the cinder volume create request may error out in > scheduling. You may also see an error when instantiating a volume from an > image if it passes the prior but then becomes out of space during writing > the image to the volume. > > I'm not sure if it's still the case, but in Havana (I see no reason for it > to change) the free space check in cinder didn't account for the difference > between promised space (the max of the volumes assigned) instead it would > literally look for free space in the output of `rados df` > > As noted above if the cluster gets to "100%" used, bad things will happen to > your VM's. The most likely case is that they all assert read-only > filesystems. (100% is a missnomer as there is a configured max % where it > will stop accepting data writes to ensure that important object replication > / maintenance can occur and have the cluster not fall over) > > On Mon, Oct 19, 2015 at 7:51 AM LOPEZ Jean-Charles <jelopez@xxxxxxxxxx> > wrote: >> >> Hi, >> >> when an OSD gets full, any write operation to the entire cluster will be >> disabled. >> >> As a result, creating a single RBD will become impossible and all VMs that >> need to write to one of their Ceph back RBDs will suffer the same pain. >> >> Usually, this ends up as a bad sorry for the VMs. >> >> The best practice is to monitor the disk space usage for the OSDs and as a >> matter of fact RHCS 1.# includes a cep old df command to do this. You can >> also use the output of the cep old report command to grab the appropriate >> info to compute it or rely on external SNMP monitoring tools to grab the >> usage details of the particular OSD disk drives. >> >> Have a great day. >> JC >> >> > On Oct 19, 2015, at 02:32, Bharath Krishna <BKrishna@xxxxxxxxxxxxxxx> >> > wrote: >> > >> > I mean cluster OSDs are physically full. >> > >> > I understand its not a pretty way to operate CEPH allowing to become >> > full, >> > but I just wanted to know the boundary condition if it becomes full. >> > >> > Will cinder create volume operation creates new volume at all or error >> > is >> > thrown at Cinder API level itself stating that no space available? >> > >> > When IO stalls, will I be able to read the data from CEPH cluster I.e >> > can >> > I still read data from existing volumes created from CEPH cluster? >> > >> > Thanks for the quick reply. >> > >> > Regards >> > M Bharath Krishna >> > >> > On 10/19/15, 2:51 PM, "Jan Schermer" <jan@xxxxxxxxxxx> wrote: >> > >> >> Do you mean when the CEPH cluster (OSDs) is physically full or when the >> >> quota is reached? >> >> >> >> If CEPH becomes full it just stalls all IO (maybe just write IO, but >> >> effectively same thing) - not pretty and you must never ever let it >> >> become full. >> >> >> >> Jan >> >> >> >> >> >>> On 19 Oct 2015, at 11:15, Bharath Krishna <BKrishna@xxxxxxxxxxxxxxx> >> >>> wrote: >> >>> >> >>> Hi >> >>> >> >>> What happens when Cinder service with CEPH backend storage cluster >> >>> capacity is FULL? >> >>> >> >>> What would be the out come of new cinder create volume request? >> >>> >> >>> Will volume be created with space not available for use or an error >> >>> thrown from Cinder API stating no space available for new volume. >> >>> >> >>> I could not try this in my environment and fill up the cluster. >> >>> >> >>> Please reply if you have ever tried and tested this. >> >>> >> >>> Thank you. >> >>> >> >>> Regards, >> >>> M Bharath Krishna >> >>> _______________________________________________ >> >>> ceph-users mailing list >> >>> ceph-users@xxxxxxxxxxxxxx >> >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> >> > >> > _______________________________________________ >> > ceph-users mailing list >> > ceph-users@xxxxxxxxxxxxxx >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@xxxxxxxxxxxxxx >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > -- > > -- > > Andrew Woodward > > Mirantis > > Fuel Community Ambassador > > Ceph Community > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com