Re: Pause i/o from time to time

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I have examined logs.
Yes, first time it can be scrubbing. It repaired some self. 

I had 2 servers before first problem: one dedicated for osd (osd.0), and second - with osd and websites (osd.1).
After problem I add third server - dedicated for osd (osd.2) and call
ceph osd set out osd.1 for replace data.

In ceph -s i saw normal replacing process and all work good about 5-7 hours.
Then I have many misdirected records (few hundreds per second):
osd.0 [WRN] client.359671  misdirected client.359671.1:220843 pg 2.3ae744c0 to osd.0 not [2,0] in e1040/1040
and errors in i/o operations.

Now I have about 20GB ceph logs with this errors. (I don't work with cluster now - I copy out all data on hdd and work from hdd).

Is any way have local software raid1 with ceph rbd and local image (for work when ceph fail or work slow by any reason).
I tried mdadm but it work bad - server hang up every few hours.

> You could be suffering from a known, but unfixed issue [1] where spindle contention from scrub and deep-scrub cause periodic stalls in RBD. You can try to disable scrub and deep-scrub with:
> 
> # ceph osd set noscrub
> # ceph osd set nodeep-scrub
> 
> If your problem stops, Issue #6278 is likely the cause. To re-enable scrub and deep-scrub:
> 
> # ceph osd unset noscrub
> # ceph osd unset nodeep-scrub
> 
> Because you seem to only have two OSDs, you may also be saturating your disks even without scrub or deep-scrub.
> 
> http://tracker.ceph.com/issues/6278
> 
> Cheers,
> Mike Dawson
> 
> 
> On 9/16/2013 12:30 PM, Timofey wrote:
>> I use ceph for HA-cluster.
>> Some time ceph rbd go to have pause in work (stop i/o operations). Sometime it can be when one of OSD slow response to requests. Sometime it can be my mistake (xfs_freeze -f for one of OSD-drive).
>> I have 2 storage servers with one osd on each. This pauses can be few minutes.
>> 
>> 1. Is any settings for fast change primary osd if current osd work bad (slow, don't response).
>> 2. Can I use ceph-rbd in software raid-array with local drive, for use local drive instead of ceph if ceph cluster fail?
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@xxxxxxxxxxxxxx
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> 

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux