On Mon, Sep 14, 2015 at 04:39:33PM +0530, Parav Pandit wrote: > 1. How does the % of resource, is different than absolute number? With > rest of the cgroups systems we define absolute number at most places > to my knowledge. There isn't really much choice if the abstraction is a bundle of all resources. You can't use an absolute number unless every possible hardware limited resource is defined, which doesn't seem smart to me either. It is not abstract enough, and doesn't match our universe of hardware very well. > 2. bytes of kernel memory for RDMA structures > One QP of one vendor might consume X bytes and other Y bytes. How does > the application knows how much memory to give. I don't see this distinction being useful at such a fine granularity where the control side needs to distinguish between 1 and 2 QPs. The majority use for control groups has been along with containers to prevent a container for exhausting resources in a way that impacts another. In that use model limiting each container to N MB of kernel memory makes it straightforward to reason about resource exhaustion in a multi-tennant environment. We have other controllers that do this, just more indirectly (ie limiting the number of inotifies, or the number of fds indirectly cap kernel memory consumption) ie Presumably some fairly small limitation like 10MB is enough for most non-MPI jobs. > Application doing 100 QP allocation, still within limit of memory of > cgroup leaves other applications without any QP. No, if the HW has a fixed QP pool then it would hit #1 above. Both are active at once. For example you'd say a container cannot use more than 10% of the device's hardware resources, or more than 10MB of kernel memory. If on an mlx card, you probably hit the 10% of QP resources first. If on an qib card there is no HW QP pool (well, almost, QPNs are always limited), so you'd hit the memory limit instead. In either case, we don't want to see a container able to exhaust either all of kernel memory or all of the HW resources to deny other containers. If you have a non-container use case in mind I'd be curious to hear it.. > I don't see a point of memory footprint based scheme, as memory limits > are well addressed by more smarter memory controller anyway. I don't thing #1 is controlled but another controller. This is long lived kernel-side memory allocations to support RDMA resource allocation - we certainly have nothing in the rdma layer that is tracking this stuff. > If the hardware vendor defines the resource pool without saying its > resource QP or MR, how would actually management/control point can > decide what should be controlled to what limit? In the kernel each HW driver has to be involved to declare what it's hardware resource limits are. In user space, it is just a simple limiter knob to prevent resource exhaustion. UAPI wise, nobdy has to care if the limit is actually # of QPs or something else. Jason -- To unsubscribe from this list: send the line "unsubscribe linux-doc" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html