Team,
I am supporting Oracle MOS note 1354980.1, which covers tuning clients
for RMAN backup to the ZFS Storage Appliance. One of the tuning
recommendations is to change sunrpc.tcp_slot_table_entries from the
default (16) to 128 to open up the number of concurrent I/O we can get
per client mount point. This is presumed good for general-purpose
kernel NFS application traffic to the ZFS Storage Appliance. I recently
received the following comment regarding the efficacy of the
sunrpc.tcp_slot_table_entries tune:
"In most cases, the parameter "sunrpc.tcp_slot_table_entries" can not be
set even if applying int onto /etc/sysctl.conf although this document
says users should do so.
Because, the parameter is appeared after sunrpc.ko module is loaded(=NFS
service is started), and sysctl was executed before starting NFS service."
I'd like to find out how to tell if the tune is actually in play for the
running kernel and if there is a difference in what is reported /proc
compared to what is running in core. Could anyone on the alias suggest
how to validate if the aforementioned comment is relevant for the Linux
kernel I am running with? I am familiar with using mdb on Solaris to
check what values the Solaris kernel is running with, so if there is a
Linux equivalent, or another way to do this sort of thing with Linux,
please let me know.
Thanks,
Jeff
--
To unsubscribe from this list: send the line "unsubscribe linux-nfs" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html