On Wed, 19 Sep 2012, Tren Blackburn wrote: > On Wed, Sep 19, 2012 at 1:52 PM, Gregory Farnum <greg@xxxxxxxxxxx> wrote: > > On Wed, Sep 19, 2012 at 1:48 PM, Sage Weil <sage@xxxxxxxxxxx> wrote: > >> On Wed, 19 Sep 2012, Tren Blackburn wrote: > >>> Hey List; > >>> > >>> I'm in the process of rsyncing in about 7TB of data to Ceph across > >>> approximately 58565475 files (okay, so I guess that's not so > >>> approximate). It's only managed to copy a small portion of this so far > >>> (about 35GB) and the server that currently is the mds master shows the > >>> ceph-mds process closing in on 2GB of RAM used. I'm getting a little > >>> nervous about this steep increase. > >>> > >>> fern ceph # ps wwaux | grep ceph-mds > >>> root 29943 3.7 0.9 2473884 1915468 ? Ssl 08:42 11:22 > >>> /usr/bin/ceph-mds -i 1 --pid-file /var/run/ceph/mds.1.pid -c > >>> /etc/ceph/ceph.conf > >>> > >>> Thank in advance! > >> > >> That's not necessarily a problem. The mds memory usage is controlled via > >> the 'mds cache size' knob which defaults to 100,000. You might try > >> halving that and restarting the MDS. > > > > Last time we counted those inodes were ~1KB each, so getting 2GB of > > memory used is an awful lot. > > > > Tren, what does your tree look like ? are there very large directories? > > Greg: It's difficult to tell you that. I'm rsyncing 2 volumes from our > filers. Each base directory on each filer mount has approximate 213 > directories, and then each directory under that has approximately > anywhere from 3000 - 5000 directories (very loose approximation here, > 850,000 directories per filer mount), and then each of those > directories contains files. > > We have many many files here. We're doing this to see how CephFS > handles lots of files. We are coming from MooseFS which its master > metalogger process eats lots of ram, so we're hoping that Ceph is a > bit lighter on us. > > Sage: The memory the MDS is using is only a cache? There should be no > problem restarting the MDS server while activity is going on? I should > probably change the limit for the non-active MDS servers first, and > then the active one and hope it fails over cleanly? Right. Restarting ceph-mds will do the trick. You shouldn't have to adjust max_mds (leave it at 1). Just make sure it doesn't fail over to another ceph-mds daemon that was started before you made the config change. There is some pending work to reduce the memory footprint for in-memory inodes; they currently include a zillion fields used for dirty updates that aren't needed for clean/cached metadata. Fixing that is a boring but tedious process of moving them into an auxilliary structure and will happen sometime soonish. :) sage -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html