Default stack size shouldn't matter. At least it's not an issue on a kernel with over-commit turned on (default). Most threads / apps never use that many stack frames (in fact they use a fraction of that), thus the kernel doesn't bother allocating the pages to it. My bet is on some other resource. On 7/23/14, 3:22 PM, Vit Yenukas wrote: > Just some fun fact pertaining to the resources consumption during startup sequence - > we've ran out of memory on a 72-disk server with 256GB RAM during the startup. > ceph-osd dies with 'can not fork' and cores. There were in excess of 40 thousands > threads when this began to happen. With default thread stack size being 8MB, no wonder :) > Note that this was in an experimental setup with just one node, so all OSDs peering happens on the same host. > Just for heck of it, I reduced the number of OSDs by two (to 36 OSDs) by setting up a soft RAID-0 for each disk pair. > This worked after some tweaking of udev rules (that ignore 'md' block devs). I'm not sure if we're going to see > the same problem with real cluster (18 such 72-disk nodes), with EC 9-3. > Also, not sure if reducing user proc stack to 4MB would be a good idea. > > On 07/22/2014 08:08 PM, Gregory Farnum wrote: > >> On Tue, Jul 22, 2014 at 6:19 AM, Wido den Hollander <wido@xxxxxxxx> wrote: >>> Hi, >>> >>> Currently on Ubuntu with Upstart when you invoke a restart like this: >>> >>> $ sudo restart ceph-osd-all >>> >>> It will restart all OSDs at once, which can increase the load on the system >>> a quite a bit. >>> >>> It's better to restart all OSDs by restarting them one by one: >>> >>> $ sudo ceph restart ceph-osd id=X >>> >>> But you then have to figure out all the IDs by doing a find in >>> /var/lib/ceph/osd and that's more manual work. >>> >>> I'm thinking of patching the init scripts which allows something like this: >>> >>> $ sudo restart ceph-osd-all delay=180 >>> >>> It then waits 180 seconds between each OSD restart making the proces even >>> smoother. >>> >>> I know there are currently sysvinit, upstart and systemd scripts, so it has >>> to be implemented on various places, but how does the general idea sound? >> That sounds like a good idea to me. I presume you're meaning to >> actually delay the restarts, not just turning them on, so that the >> daemons all remain alive (that's what it sounds like to me here, just >> wanted to clarify). >> -Greg > -- > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html