Re: ceph OSD with 95% full

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Thanks for detail...
When an OSD is 95% full, then that specific OSD's write IO blocked.

Thanks
Swami

On Tue, Jul 19, 2016 at 3:07 PM, Christian Balzer <chibi@xxxxxxx> wrote:
>
> Hello,
>
> On Tue, 19 Jul 2016 14:23:32 +0530 M Ranga Swami Reddy wrote:
>
>> >> Using ceph cluster with 100+ OSDs and cluster is filled with 60% data.
>> >> One of the OSD is 95% full.
>> >> If an OSD is 95% full, is it impact the any storage operation? Is this
>> >> impacts on VM/Instance?
>>
>> >Yes, one OSD will impact whole cluster. It will block write operations to the cluster
>>
>> Thanks for clarification. Really?? Is this(OSD 95%) full designed to
>> block write I/O of ceph cluster?
>>
> Really.
> To be more precise, any I/O that touches any PG on that OSD will block.
> So with a sufficiently large cluster you may have some, few, I/Os still go
> through as they don't use that OSD at all.
>
> That's why:
>
> 1. Ceph has the near-full warning (which of course may need to be
> adjusted to correctly reflect things, especially with smaller clusters).
> Once you get that warning, you NEED to take action immediately.
>
> 2. You want to graph the space utilization of all your OSDs with something
> like graphite. That allows you to spot trends of uneven data distribution
> early and thus react early to it.
> I re-weight (CRUSH re-weight, as this is permanent and my clusters aren't
> growing frequently) OSDs so they they are at least within 10% of each
> other.
>
> Christian
>> Because I have around 251 OSDs out which one OSD is 95% full, but
>> other 250 OSDs not in near full also...
>>
>> Thanks
>> Swami
>>
>>
>> On Tue, Jul 19, 2016 at 2:17 PM, Henrik Korkuc <lists@xxxxxxxxx> wrote:
>> > On 16-07-19 11:44, M Ranga Swami Reddy wrote:
>> >>
>> >> Hi,
>> >> Using ceph cluster with 100+ OSDs and cluster is filled with 60% data.
>> >> One of the OSD is 95% full.
>> >> If an OSD is 95% full, is it impact the any storage operation? Is this
>> >> impacts on VM/Instance?
>> >
>> > Yes, one OSD will impact whole cluster. It will block write operations to
>> > the cluster
>> >>
>> >> Immediately I have reduced the OSD weight, which was filled with 95 %
>> >> data. After re-weight, data rebalanaced and OSD came to normal state
>> >> (ie < 80%) with 1 hour time frame.
>> >>
>> >>
>> >> Thanks
>> >> Swami
>> >> _______________________________________________
>> >> ceph-users mailing list
>> >> ceph-users@xxxxxxxxxxxxxx
>> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >
>> >
>> >
>> > _______________________________________________
>> > ceph-users mailing list
>> > ceph-users@xxxxxxxxxxxxxx
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@xxxxxxxxxxxxxx
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
>
> --
> Christian Balzer        Network/Systems Engineer
> chibi@xxxxxxx           Global OnLine Japan/Rakuten Communications
> http://www.gol.com/
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux