Re: ceph OSD with 95% full

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



+1 .. I agree....

Thanks
Swami

On Tue, Jul 19, 2016 at 4:57 PM, Lionel Bouton <lionel+ceph@xxxxxxxxxxx> wrote:
> Hi,
>
> On 19/07/2016 13:06, Wido den Hollander wrote:
>>> Op 19 juli 2016 om 12:37 schreef M Ranga Swami Reddy <swamireddy@xxxxxxxxx>:
>>>
>>>
>>> Thanks for the correction...so even one OSD reaches to 95% full, the
>>> total ceph cluster IO (R/W) will be blocked...Ideally read IO should
>>> work...
>> That should be a config option, since reading while writes still block is also a danger. Multiple clients could read the same object, perform a in-memory change and their write will block.
>>
>> Now, which client will 'win' after the full flag has been removed?
>>
>> That could lead to data corruption.
>
> If it did, the clients would be broken as normal usage (without writes
> being blocked) doesn't prevent multiple clients from reading the same
> data and trying to write at the same time. So if multiple writes (I
> suppose on the same data blocks) are possibly waiting the order in which
> they are performed *must not* matter in your system. The alternative is
> to prevent simultaneous write accesses from multiple clients (this is
> how non-cluster filesystems must be configured on top of Ceph/RBD, they
> must even be prevented from read-only accessing an already mounted fs).
>
>>
>> Just make sure you have proper monitoring on your Ceph cluster. At nearfull it goes into WARN and you should act on that.
>
>
> +1 : monitoring is not an option.
>
> Lionel
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux