Search Postgresql Archives

Re: [HACKERS] Bottlenecks with large number of relation segment files

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



(2013/08/05 21:23), Tom Lane wrote:
> Andres Freund <andres@xxxxxxxxxxxxxxx> writes:
>> ... Also, there are global
>> limits to the amount of filehandles that can simultaneously opened on a
>> system.
> 
> Yeah.  Raising max_files_per_process puts you at serious risk that
> everything else on the box will start falling over for lack of available
> FD slots. 
Is it Really? When I use hadoop like NOSQL storage, I set large number of FD.
Actually, Hadoop Wiki is writing following.

http://wiki.apache.org/hadoop/TooManyOpenFiles
> Too Many Open Files
> 
> You can see this on Linux machines in client-side applications, server code or even in test runs. 
> It is caused by per-process limits on the number of files that a single user/process can have open, which was introduced in the 2.6.27 kernel. The default value, 128, was chosen because "that should be enough".
> 
> In Hadoop, it isn't.
~
> ulimit -n 8192

Regards,
--
Mitsumasa KONDO
NTT Open Source Software Center


-- 
Sent via pgsql-general mailing list (pgsql-general@xxxxxxxxxxxxxx)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-general




[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[Index of Archives]     [Postgresql Jobs]     [Postgresql Admin]     [Postgresql Performance]     [Linux Clusters]     [PHP Home]     [PHP on Windows]     [Kernel Newbies]     [PHP Classes]     [PHP Books]     [PHP Databases]     [Postgresql & PHP]     [Yosemite]
  Powered by Linux