Re: ceph mon eating lots of memory after upgrade0.94.2 to 0.94.9

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



hi Corin. We run latest hammer on CentOS7.2, with 3 mons and have not
seen this problem. I'm not sure if there are any other possible
differences between the healthy nodes and the one that has excessive
consumption of memory? thx will

On Fri, Nov 18, 2016 at 6:35 PM, Corin Langosch
<corin.langosch@xxxxxxxxxxx> wrote:
> Hi,
>
> about 2 weeks ago I upgraded a rather small cluster from ceph 0.94.2 to
> 0.94.9. The upgrade went fine, the cluster is running stable. But I just
> noticed that one monitor is already eating 20 GB of memory, growing slowly
> over time. The other 2 mons look fine. The disk space used by the
> problematic mon looks fine too.
>
> ceph -w
>     cluster 4ac0e21b-6ea2-4ac7-8114-122bd9ba55d6
>      health HEALTH_OK
>      monmap e5: 3 mons at
> {a=10.0.0.5:6789/0,b=10.0.0.6:6789/0,c=10.0.0.7:6789/0}
>             election epoch 856, quorum 0,1,2 a,b,c
>      osdmap e27838: 19 osds: 9 up, 9 in
>       pgmap v104865438: 4096 pgs, 1 pools, 3225 GB data, 809 kobjects
>             6463 GB used, 1955 GB / 8419 GB avail
>                 4096 active+clean
>   client io 1484 kB/s rd, 14217 kB/s wr, 2026 op/s
>
>
> ps aux | grep ceph
> root      5958  1.2 15.7 21194296 20821228 ?   Sl   Nov05 238:33
> /usr/bin/ceph-mon -i c --pid-file /var/run/ceph/mon.c.pid -c
> /etc/ceph/ceph.conf --cluster ceph
> root      7457 38.5  0.3 1660028 509008 ?      Ssl  Nov05 7125:00
> /usr/bin/ceph-osd -i 18 --pid-file /var/run/ceph/osd.18.pid -c
> /etc/ceph/ceph.conf --cluster ceph
> root      7981 25.8  0.4 1661064 543684 ?      Ssl  Nov05 4775:44
> /usr/bin/ceph-osd -i 9 --pid-file /var/run/ceph/osd.9.pid -c
> /etc/ceph/ceph.conf --cluster ceph
>
>
> ps aux | grep ceph
> root     12704  2.0  0.1 532184 93468 ?        Sl   Nov05 374:16
> /usr/bin/ceph-mon -i a --pid-file /var/run/ceph/mon.a.pid -c
> /etc/ceph/ceph.conf --cluster ceph
> root     14587 32.3  0.8 1682720 581752 ?      Ssl  Nov05 5970:53
> /usr/bin/ceph-osd -i 16 --pid-file /var/run/ceph/osd.16.pid -c
> /etc/ceph/ceph.conf --cluster ceph
> root     14919 28.8  0.7 1680144 526052 ?      Ssl  Nov05 5328:17
> /usr/bin/ceph-osd -i 3 --pid-file /var/run/ceph/osd.3.pid -c
> /etc/ceph/ceph.conf --cluster ceph
>
>
> ps aux | grep ceph
> root     19114  1.3  0.2 508000 157724 ?       Sl   Nov05 256:53
> /usr/bin/ceph-mon -i b --pid-file /var/run/ceph/mon.b.pid -c
> /etc/ceph/ceph.conf --cluster ceph
> root     20155 34.3  0.7 1652784 516072 ?      Ssl  Nov05 6341:37
> /usr/bin/ceph-osd -i 17 --pid-file /var/run/ceph/osd.17.pid -c
> /etc/ceph/ceph.conf --cluster ceph
> root     20508 33.2  0.7 1666276 510496 ?      Ssl  Nov05 6135:04
> /usr/bin/ceph-osd -i 6 --pid-file /var/run/ceph/osd.6.pid -c
> /etc/ceph/ceph.conf --cluster ceph
>
>
> dpkg -l | grep ceph
> ii  ceph                                 0.94.9-1trusty
> amd64        distributed storage and file system
> ii  ceph-common                          0.94.9-1trusty
> amd64        common utilities to mount and interact with a ceph storage
> cluster
> ii  ceph-fs-common                       0.94.9-1trusty
> amd64        common utilities to mount and interact with a ceph file system
> ii  ceph-fuse                            0.94.9-1trusty
> amd64        FUSE-based client for the Ceph distributed file system
> ii  ceph-mds                             0.94.9-1trusty
> amd64        metadata server for the ceph distributed file system
> ii  ceph-test                            0.94.9-1trusty
> amd64        Ceph test and benchmarking tools
> ii  libcephfs1                           0.94.9-1trusty
> amd64        Ceph distributed file system client library
> ii  python-cephfs                        0.94.9-1trusty
> amd64        Python libraries for the Ceph libcephfs library
>
> /var/lib/ceph/mon/ceph-c# du -h
> 108K    ./auth
> 88K     ./auth_gv
> 2.1M    ./logm
> 2.0M    ./logm_gv
> 24K     ./mdsmap
> 4.0K    ./mdsmap_gv
> 4.0K    ./mkfs
> 40K     ./monmap
> 20K     ./monmap_gv
> 4.2M    ./osdmap
> 15M     ./osdmap_full
> 4.2M    ./osdmap_gv
> 13M     ./pgmap
> 2.0M    ./pgmap_gv
> 201M    ./store.db
> 243M    .
>
> Is this a known bug? Any workaround or fix available?
>
> Thanks
> Corin
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux