Re: e: mon memory issue

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I try with radosgw and it's reporting very nice output from valgrind,
but still nothing from mon.

desc: (none)
cmd: /usr/bin/ceph-mon -i 0 --pid-file /var/run/ceph/mon.0.pid -c
/etc/ceph/ceph.conf -f
time_unit: i
#-----------
snapshot=0
#-----------
time=0
mem_heap_B=0
mem_heap_extra_B=0
mem_stacks_B=0
heap_tree=empty

On Wed, Sep 5, 2012 at 8:44 PM, Sławomir Skowron <szibis@xxxxxxxxx> wrote:
> On Wed, Sep 5, 2012 at 5:51 PM, Sage Weil <sage@xxxxxxxxxxx> wrote:
>> On Wed, 5 Sep 2012, S?awomir Skowron wrote:
>>> Unfortunately here is the problem in my Ubuntu 12.04.1
>>>
>>> --9399-- You may be able to write your own handler.
>>> --9399-- Read the file README_MISSING_SYSCALL_OR_IOCTL.
>>> --9399-- Nevertheless we consider this a bug.  Please report
>>> --9399-- it at http://valgrind.org/support/bug_reports.html.
>>> ==9399== Warning: noted but unhandled ioctl 0x9408 with no size/direction hints
>>> ==9399==    This could cause spurious value errors to appear.
>>> ==9399==    See README_MISSING_SYSCALL_OR_IOCTL for guidance on
>>> writing a proper wrapper.
>>> --9399-- WARNING: unhandled syscall: 306
>>> --9399-- You may be able to write your own handler.
>>> --9399-- Read the file README_MISSING_SYSCALL_OR_IOCTL.
>>> --9399-- Nevertheless we consider this a bug.  Please report
>>> --9399-- it at http://valgrind.org/support/bug_reports.html.
>>> ==9399== Warning: noted but unhandled ioctl 0x9408 with no size/direction hints
>>> ==9399==    This could cause spurious value errors to appear.
>>> ==9399==    See README_MISSING_SYSCALL_OR_IOCTL for guidance on
>>> writing a proper wrapper.
>>
>> These are harmless; it just doesn't recognize syncfs(2) or one of the
>> ioctls, but everything else works.
>>
>>> ^C2012-09-05 09:13:18.660048 a964700 -1 mon.0@0(leader) e4 *** Got
>>> Signal Interrupt ***
>>> ==9399==
>>
>> Did you hit control-c?
>>
>> If you leave it running it should gather the memory utilization info we
>> need...
>
> Yes it's running now, and  i will see tomorrow how much memory mon consumes.
>
>>
>> sage
>>
>>>
>>>
>>> On Wed, Sep 5, 2012 at 5:32 AM, Sage Weil <sage@xxxxxxxxxxx> wrote:
>>> > On Tue, 4 Sep 2012, S?awomir Skowron wrote:
>>> >> Valgrind returns nothing.
>>> >>
>>> >> valgrind --tool=massif --log-file=ceph_mon_valgrind ceph-mon -i 0 > log.txt
>>> >
>>> > The fork is probably confusing it.  I usually pass -f to ceph-mon (or
>>> > ceph-osd etc) to keep it in the foreground.  Can you give that a go?
>>> > e.g.,
>>> >
>>> >         valgrind --tool-massif ceph-mon -i 0 -f &
>>> >
>>> > and watch for the massif.out.$pid file.
>>> >
>>> > Thanks!
>>> > sage
>>> >
>>> >
>>> >>
>>> >> ==30491== Massif, a heap profiler
>>> >> ==30491== Copyright (C) 2003-2011, and GNU GPL'd, by Nicholas Nethercote
>>> >> ==30491== Using Valgrind-3.7.0 and LibVEX; rerun with -h for copyright info
>>> >> ==30491== Command: ceph-mon -i 0
>>> >> ==30491== Parent PID: 4013
>>> >> ==30491==
>>> >> ==30491==
>>> >>
>>> >> cat massif.out.26201
>>> >> desc: (none)
>>> >> cmd: ceph-mon -i 0
>>> >> time_unit: i
>>> >> #-----------
>>> >> snapshot=0
>>> >> #-----------
>>> >> time=0
>>> >> mem_heap_B=0
>>> >> mem_heap_extra_B=0
>>> >> mem_stacks_B=0
>>> >> heap_tree=empty
>>> >>
>>> >> What i have done wrong ??
>>> >>
>>> >> On Fri, Aug 31, 2012 at 8:34 PM, S?awomir Skowron <szibis@xxxxxxxxx> wrote:
>>> >> > I have this problem too. My mon's in 0.48.1 cluster have 10GB RAM
>>> >> > each, with 78 osd, and 2k request per minute (max) in radosgw.
>>> >> >
>>> >> > Now i have run one via valgrind. I will send output when mon grow up.
>>> >> >
>>> >> > On Fri, Aug 31, 2012 at 6:03 PM, Sage Weil <sage@xxxxxxxxxxx> wrote:
>>> >> >> On Fri, 31 Aug 2012, Xiaopong Tran wrote:
>>> >> >>
>>> >> >>> Hi,
>>> >> >>>
>>> >> >>> Is there any known memory issue with mon? We have 3 mons running, and
>>> >> >>> on keeps on crashing after 2 or 3 days, and I think it's because mon
>>> >> >>> sucks up all memory.
>>> >> >>>
>>> >> >>> Here's mon after starting for 10 minutes:
>>> >> >>>
>>> >> >>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
>>> >> >>> 13700 root      20   0  163m  32m 3712 S   4.3  0.1   0:05.15 ceph-mon
>>> >> >>>  2595 root      20   0 1672m 523m    0 S   1.7  1.6 954:33.56 ceph-osd
>>> >> >>>  1941 root      20   0 1292m 220m    0 S   0.7  0.7 946:40.69 ceph-osd
>>> >> >>>  2316 root      20   0 1169m 198m    0 S   0.7  0.6 420:26.74 ceph-osd
>>> >> >>>  2395 root      20   0 1149m 184m    0 S   0.7  0.6 364:29.08 ceph-osd
>>> >> >>>  2487 root      20   0 1354m 373m    0 S   0.7  1.2 401:13.97 ceph-osd
>>> >> >>>   235 root      20   0     0    0    0 S   0.3  0.0   0:37.68 kworker/4:1
>>> >> >>>  1304 root      20   0     0    0    0 S   0.3  0.0   0:00.16 jbd2/sda3-8
>>> >> >>>  1327 root      20   0     0    0    0 S   0.3  0.0  13:07.00 xfsaild/sdf1
>>> >> >>>  2011 root      20   0 1240m 177m    0 S   0.3  0.6 411:52.91 ceph-osd
>>> >> >>>  2153 root      20   0 1095m 166m    0 S   0.3  0.5 370:56.01 ceph-osd
>>> >> >>>  2725 root      20   0 1214m 186m    0 S   0.3  0.6 378:16.59 ceph-osd
>>> >> >>>
>>> >> >>> Here's the memory situation of mon on another machine, after mon has
>>> >> >>> been running for 3 hours:
>>> >> >>>
>>> >> >>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
>>> >> >>>  1716 root      20   0 1923m 1.6g 4028 S   7.6  5.2   8:45.82 ceph-mon
>>> >> >>>  1923 root      20   0  774m 138m 5052 S   0.7  0.4   1:28.56 ceph-osd
>>> >> >>>  2114 root      20   0  836m 143m 4864 S   0.7  0.4   1:20.14 ceph-osd
>>> >> >>>  2304 root      20   0  863m 176m 4988 S   0.7  0.5   1:13.30 ceph-osd
>>> >> >>>  2578 root      20   0  823m 150m 5056 S   0.7  0.5   1:24.55 ceph-osd
>>> >> >>>  2781 root      20   0  819m 131m 4900 S   0.7  0.4   1:12.14 ceph-osd
>>> >> >>>  2995 root      20   0  863m 179m 5024 S   0.7  0.6   1:41.96 ceph-osd
>>> >> >>>  3474 root      20   0  888m 208m 5608 S   0.7  0.6   7:08.08 ceph-osd
>>> >> >>>  1228 root      20   0     0    0    0 S   0.3  0.0   0:07.01 jbd2/sda3-8
>>> >> >>>  1853 root      20   0  859m 176m 4820 S   0.3  0.5   1:17.01 ceph-osd
>>> >> >>>  3373 root      20   0  789m 118m 4916 S   0.3  0.4   1:06.26 ceph-osd
>>> >> >>>
>>> >> >>> And here is the situation on a third node, mon has been running
>>> >> >>> for over a week:
>>> >> >>>
>>> >> >>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
>>> >> >>>  1717 root      20   0 68.8g  26g 2044 S  91.5 84.1   9220:40 ceph-mon
>>> >> >>>  1986 root      20   0 1281m 226m    0 S   1.7  0.7   1225:28 ceph-osd
>>> >> >>>  2196 root      20   0 1501m 538m    0 S   1.0  1.7   1221:54 ceph-osd
>>> >> >>>  2266 root      20   0 1121m 176m    0 S   0.7  0.5 399:23.70 ceph-osd
>>> >> >>>  2056 root      20   0 1072m 167m    0 S   0.3  0.5 403:49.76 ceph-osd
>>> >> >>>  2126 root      20   0 1412m 458m    0 S   0.3  1.4   1215:48 ceph-osd
>>> >> >>>  2337 root      20   0 1128m 188m    0 S   0.3  0.6 408:31.88 ceph-osd
>>> >> >>>
>>> >> >>> So, after a while, sooner or later, mon is going to crash, just
>>> >> >>> a matter of time.
>>> >> >>>
>>> >> >>> Does anyone see anything like this? This is kinda scary.
>>> >> >>>
>>> >> >>> OS: Debian Wheezy 3.2.0-3-amd64
>>> >> >>> Ceph: 0.48argonaut (commit:c2b20ca74249892c8e5e40c12aa14446a2bf2030)
>>> >> >>
>>> >> >> Can you try with 0.48.1argonaut?
>>> >> >>
>>> >> >> If it still happens, can you run ceph-mon through massif?
>>> >> >>
>>> >> >>  valgrind --tool=massif ceph-mon -i whatever
>>> >> >>
>>> >> >> That'll generate a massif.out file (make sure it's there; you may need to
>>> >> >> specify the output file for valgrind) over time.  Once ceph-mon starts
>>> >> >> eating ram, send us a copy of the file and we can hopefully see what is
>>> >> >> leaking.
>>> >> >>
>>> >> >> Thanks!
>>> >> >> sage
>>> >> >>
>>> >> >>
>>> >> >>>
>>> >> >>> With this issue on hand, I'll have to monitor it closely and
>>> >> >>> restart mon once in a while, or I will get a crash (which is
>>> >> >>> still good enough), or a system that does not respond at
>>> >> >>> all because memory is exhausted, and the whole ceph cluster
>>> >> >>> is unreachable. We had this problem in the morning, mon on one
>>> >> >>> node exhausted the memory, none of the ceph command responds
>>> >> >>> anymore, the only thing left to do is to hard reset the node.
>>> >> >>> The whole cluster was basically done at that time.
>>> >> >>>
>>> >> >>> Here is our usage situation:
>>> >> >>>
>>> >> >>> 1) A few applications which read and write data through
>>> >> >>> librados API, we have about 20-30 connections at any one time.
>>> >> >>> So far, our apps have no such memory issue, we have been
>>> >> >>> monitoring them closely.
>>> >> >>>
>>> >> >>> 2) We have a few scripts which pull data from an old storage
>>> >> >>> system, and use the rados command to put it into ceph.
>>> >> >>> Basically, just shell script. Each rados command is run
>>> >> >>> to write one object (one file), and exit. We run about
>>> >> >>> 25 scripts simultaneously, which means at any one time,
>>> >> >>> there are at most 25 connections.
>>> >> >>>
>>> >> >>> I don't think this is a very busy system. But this
>>> >> >>> memory issue is definitely a problem for us.
>>> >> >>>
>>> >> >>> Thanks for helping.
>>> >> >>>
>>> >> >>> Xiaopong
>>> >> >>> --
>>> >> >>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>> >> >>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>>> >> >>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>> >> >>>
>>> >> >>>
>>> >> >> --
>>> >> >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>> >> >> the body of a message to majordomo@xxxxxxxxxxxxxxx
>>> >> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>> >> >
>>> >> >
>>> >> >
>>> >> > --
>>> >> > -----
>>> >> > Pozdrawiam
>>> >> >
>>> >> > S?awek "sZiBis" Skowron
>>> >>
>>> >>
>>> >>
>>> >> --
>>> >> -----
>>> >> Pozdrawiam
>>> >>
>>> >> S?awek "sZiBis" Skowron
>>> >> --
>>> >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>> >> the body of a message to majordomo@xxxxxxxxxxxxxxx
>>> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>> >>
>>> >>
>>>
>>>
>>>
>>> --
>>> -----
>>> Pozdrawiam
>>>
>>> S?awek "sZiBis" Skowron
>>>
>>>
>
>
>
> --
> -----
> Pozdrawiam
>
> Sławek "sZiBis" Skowron



-- 
-----
Pozdrawiam

Sławek "sZiBis" Skowron
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux