Re: teuthology : 70 workers need more than 8GB RAM / 2 CPUS

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Loic,

This is definitely something to keep an eye on. It's actually a bit surprising to me, though - I haven't seen ansible-playbook use any significant resources in sepia.

I wouldn't really recommend running paddles on the same host as teuthology though, to do any serious amount of testing; some teuthology tasks do use large amounts of RAM and/or CPU, and severe load issues could feasibly cause requests to time out, affecting other jobs.

That's all theory though, as I've always used separate hosts for the two services.

Zack

----- Original Message -----
> From: "Loic Dachary" <loic@xxxxxxxxxxx>
> To: "Zack Cerza" <zack@xxxxxxxxxx>, "Andrew Schoen" <aschoen@xxxxxxxxxx>
> Cc: "Ceph Development" <ceph-devel@xxxxxxxxxxxxxxx>
> Sent: Sunday, July 19, 2015 9:06:41 AM
> Subject: Re: teuthology : 70 workers need more than 8GB RAM / 2 CPUS
> 
> Hi again,
> 
> I had the same problem when 50 workers kick in at the same time. I've lowered
> the number of workers down to 25 and it went well. During a few minutes (~8
> minutes) the load average stayed around 25 (CPU bound, mainly the ansible
> playbook competing, see the screenshot of htop). But did not see any error /
> timeout. then I added 15 workers, wait for the load to go back to < 2 (10
> minutes), then 15 more (10 minutes) to get to 55.
> 
> That sound like a log of CPU used by a single playbook run. Is there a known
> way to reduce that ? If not I'll just upgrade the machine. Just want to make
> sure I'm not missing a simple solution ;-)
> 
> Cheers
> 
> On 19/07/2015 14:22, Loic Dachary wrote:
> > Hi,
> > 
> > For the record, I launched a rados suite on an idle teuthology cluster,
> > with 70 workers running on a 8GB RAM / 2 CPUS / 40GB SSD disk. The load
> > average reached 40 within a minute or two and some jobs started failing /
> > timeouting. I had pulpito running on the same machine and it failed one
> > time out of two because of the load (see the top image).
> > 
> > On friday I was able to run 70 workers because I gradually added them. The
> > load peak is when a job starts and all workers kick in a the same time.
> > 
> > Cheers
> > 
> 
> --
> Loïc Dachary, Artisan Logiciel Libre
> 
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux