Do we have any tool to monitor the OSDs usage with help of UI? Thanks Swami On Tue, Jul 19, 2016 at 6:44 PM, M Ranga Swami Reddy <swamireddy@xxxxxxxxx> wrote: > +1 .. I agree.... > > Thanks > Swami > > On Tue, Jul 19, 2016 at 4:57 PM, Lionel Bouton <lionel+ceph@xxxxxxxxxxx> wrote: >> Hi, >> >> On 19/07/2016 13:06, Wido den Hollander wrote: >>>> Op 19 juli 2016 om 12:37 schreef M Ranga Swami Reddy <swamireddy@xxxxxxxxx>: >>>> >>>> >>>> Thanks for the correction...so even one OSD reaches to 95% full, the >>>> total ceph cluster IO (R/W) will be blocked...Ideally read IO should >>>> work... >>> That should be a config option, since reading while writes still block is also a danger. Multiple clients could read the same object, perform a in-memory change and their write will block. >>> >>> Now, which client will 'win' after the full flag has been removed? >>> >>> That could lead to data corruption. >> >> If it did, the clients would be broken as normal usage (without writes >> being blocked) doesn't prevent multiple clients from reading the same >> data and trying to write at the same time. So if multiple writes (I >> suppose on the same data blocks) are possibly waiting the order in which >> they are performed *must not* matter in your system. The alternative is >> to prevent simultaneous write accesses from multiple clients (this is >> how non-cluster filesystems must be configured on top of Ceph/RBD, they >> must even be prevented from read-only accessing an already mounted fs). >> >>> >>> Just make sure you have proper monitoring on your Ceph cluster. At nearfull it goes into WARN and you should act on that. >> >> >> +1 : monitoring is not an option. >> >> Lionel _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com